var/home/core/zuul-output/0000755000175000017500000000000015145354015014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015145357533015504 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000231450015145357357020272 0ustar corecoreޕikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9GfB A}"mv?_eGbuuțx{w7ݭ7֫'e% oo/q3m^]/o?8.7oW}ʋghewx/mX,ojŻ ^Tb3b#׳:}=p7뼝ca㑔`e0I1Q!&ѱ[/o^{W-{t3_U|6 x)K#/5ΌR"ggóisR)N %emOQ/Ϋ[oa0vs68/Jʢ ܚʂ9ss3+aô٥J}{37FEbп3 FKX1QRQlrTvb)E,s)Wɀ;$#LcdHM%vz_. o~I|3j dF{ "IΩ?PF~J~ ` 17ׅwڋًM)$Fiqw7Gt7L"u 0V9c  ˹dvYļU[ Z.׿-h QZ*U1|t5wKOؾ{mk b2 ܨ;RJK!b>JR*kl|+"N'C_#a7]d]sJg;;>Yp׫,w`ɚ'd$ecwŻ^~7EpQС3DCS[Yʧ?DDS aw߾)VxX帟AB}nyи0stĈCo.:wAZ{sy:7qsWctx{}n-+ZYsI{/.Ra9XcђQ0FK@aEDO2es ׇN# ZF͹b,*YVi+$<QMGhC}^}?BqG!(8l K3T[<~6]90}(*T7siv'=k 9Q2@vN ( R['>v*;o57sp$3ncx!>t®W>]tF-iܪ%GYbaRvHa}dkD̶*';ک|s_}8yj,('GrgTZ'U鋊TqOſ * /Ijo!՟8`"j}zӲ$k3jS|C7;A)͎V.r?t\WU1ojjr<~Tq> `=tJ!aݡ=h6Yݭw}?lѹ`f_" J9w4ts7NG GGG]ҡgc⌝M b/Ζlpah E ur C&`XR JcwB~R2EL9j7e\(Uё$׿atyХ?*t5z\+`/ErVQUxMҔ&ۈt.3;eg_O ξL1KiYLizpV:C5/=v-}҅"o ']쌕|tϓX8nJ*A*%J[T2pI1Je;s_[,Ҩ38_ь ͰM0ImY/MiVJ5&jNgBt90v߁R:~U jځU~oN9xԞ~J|dݤ߯R> kH&Y``:"s ayiBq)u%'4 yܽ yW0 -i̭uJ{KưЖ@+UBj -&JO x@}DS.€>3T0|9ē7$3z^.I< )9qf e%dhy:O40n'c}c1XҸuFiƠIkaIx( +")OtZ l^Z^CQ6tffEmDφǽ{QiOENG{P;sHz"G- >+`قSᔙD'Ad ѭj( ہO r:91v|ɛr|٦/o{C Ӹ!uWȳ)gjw&+uߕt*:͵UMQrN@fYDtEYZb4-UCqK٪L.2teB ˛"ո{Gci`du듎q+;C'16FgVlWaaB)"F,u@30YQg˾_YҊŏ#_f^ TD=VAKNl4Kš4GScѦa0 J ()¾5m'p/\խX\=z,Mw˭x:qu礛WԓL!I xӤ1(5AKRVF2ɌУլ F "vuhc=JS\kkZAY`R"Hr1]%oR[^oI]${&L8<=#0yaKL: JJl r;t#H+B|ɧJiM cm)>H=l}.^\ݧM<lu Y> XH\z:dHElL(uHR0i#q%]!=t_쾋-, vW~* ^g/5n]FhNU˿oۂ6C9C7sn,kje*;iΓA7,Q)-,=1A sK|ۜLɽy]ʸEO<-YEqKzϢ \{>dDLF amKGm+`VLJsC>?5rk{-3Ss`y_C}Q v,{*)ߎ% qƦat:D=uNvdߋ{Ny[$ {ɴ6hOI']dC5`t9:GO: FmlN*:g^;T^B0$B%C6Θ%|5u=kkN2{'FEc* A>{avdt)8|mg定TN7,TEXt+`F P |ɧ<Ғ8_iqE b}$B#fethBE;1"l r  B+R6Qp%;R8P󦟶Ub-L::;Ⱦ7,VW.JE:PgXoΰUv:ΰdɆΰ (ΰ0eTUgXun[g, ׽-t!X򴱞_aM:E.Qg1DllЊE҉L ehJx{̗Uɾ?si&2"C]u$.`mjmƒVe9f6NŐsLu6fe wkىKR%f"6=rw^)'Hz }x>1yFX09'A%bDb0!i(`Z;TyֻΗ|ִ0-6dAC5t[OM91c:VJR9&ksvJ;0ɝ$krogB= FYtЩOte=?>T&O{Ll)HClba1PIFĀ":tu^}.&R*!^pHPQuSVO$.KMb.:DK>WtWǭKv4@Va3"a`R@gbu%_J5Ґ 3?lm$K/$s_. WM]̍"W%`lO2-"ew@E=! I,($F{ձ7*Oy 6EK( EF #31J8mN .TTF9㕴/5~RxCe,&v3,JE- ZF5%Da,Gܠ*qI@qlG6s푻jÝ$ >8ȕ$eZ1j[h0SH,qf<"${/ksBK}xnwDb%M6:K<~̓9*u᛹Q{FЖt~6S#G1(zr6<ߜ!?U\(0EmG4 4c~J~]ps/9܎ms4gZY-07`-Id,9õ԰t+-b[uemNi_󈛥^g+!SKq<>78NBx;c4<ニ)H .Pd^cR^p_G+E--ۥ_F]a|v@|3p%kzh|k*BBRib\J3Yn|뇱[FfP%M:<`pz?]6laz5`ZQs{>3ư_o%oU׆]YLz_s߭AF'is^_&uUm$[[5HI4QCZ5!N&D[uiXk&2Bg&Ս7_/6v_cd쿽d@eU XyX2z>g8:.⺻h()&nO5YE\1t7aSyFxPV19 ĕi%K"IcB j>Pm[E[^oHmmU̸nG pHKZ{{Qo}i¿Xc\]e1e,5`te.5Hhao<[50wMUF􀍠PV?Yg"ź)\3mf|ܔMUiU|Ym! #'ukMmQ9Blm]TO1ba.XW x6ܠ9[v35H;-]Um4mMrW-k#~fؤϋu_j*^Wj^qM `-Pk.@%=X#|ۡb1lKcj$׋bKv[~"N jS4HOkeF3LPyi︅iWk! cAnxu6<7cp?WN $?X3l(?  'Z! ,Z.maO_Bk/m~ޖ(<qRfR"Au\PmLZ"twpuJ` mvf+T!6Ѓjw1ncuwo':o gSPC=]U҅yY9 &K<-na'Xk,P4+`Þ/lX/bjFO.= w ?>ȑ3n߿z,t s5Z/ Clo-` z?a~b mzkC zFȏ>1k*Dls6vP9hS  ehC.3 @6ijvUuBY hBnb[ Fr#D7ćlA!:X lYE>#0JvʈɌ|\u,'Y˲.,;oOwoj-25Hݻ7 li0bSlbw=IsxhRbd+I]Y]JP}@.供SЃ??w w@KvKts[TSa /ZaDžPAEư07>~w3n:U/.P珀Yaٳ5Ʈ]խ4 ~fh.8C>n@T%W?%TbzK-6cb:XeGL`'žeVVޖ~;BLv[n|viPjbMeO?!hEfޮ])4 ?KN1o<]0Bg9lldXuT ʑ!Iu2ʌnB5*<^I^~G;Ja߄bHȌsK+D"̽E/"Icƀsu0,gy(&TI{ U܋N5 l͖h"褁lm *#n/Q!m b0X3i)\IN˭% Y&cKoG w 9pM^WϋQf7s#bd+SDL ,FZ<1Kx&C!{P|Ռr,* ] O;*X]Eg,5,ouZm8pnglVj!p2֬uT[QyB402|2d5K: `Bcz|Rxxl3{c` 1nhJzQHv?hbºܞz=73qSO0}Dc D]ͺjgw07'㤸z YJ\Hb9Ɖ„2Hi{(2HFE?*w*hy4ޙM^٫wF(p]EwQzr*! 5F XrO7E[!gJ^.a&HߣaaQÝ$_vyz4}0!yܒ栒޹a% Ŋ X!cJ!A\ ?E\R1 q/rJjd A4y4c+bQ̘TT!kw/nb͵FcRG0xeOӜ9wWLeUc.)#/udx4~kƊզ7.LJ5jRS}_D U4x[c) ,`̔Dvckk5Ťã0le۞]o~oW(91ݧ$uxp/Cq6Un9%Z`.hI"!A6:uQզ}@j=Mo<}nYUw1Xw:]e/sm lˣaVۤkĨdԖ)RtS2 "E I"{;ōCb{yex&Td >@).p$`XKxnX~E膂Og\IGֻq<-uˮ◶>waPcPw3``m- } vS¢=j=1 W=&;JW(7b ?Q.|K,ϩ3g)D͵Q5PBj(h<[rqTɈjM-y͢FY~p_~O5-֠kDNTͷItI1mk"@$AǏ}%S5<`d+0o,AրcbvJ2O`gA2Ȏp@Z#"U4Xk1G;7#m eji'ĒGIqB//(O &1I;svHd=mJW~ړUCOīpAiB^MP=MQ`=JB!"]b6Ƞi]ItЀ'Vf:yo=K˞r:( n72-˒#K9T\aVܩO "^OF1%e"xm뻱~0GBeFO0ޑ]w(zM6j\v00ׅYɓHڦd%NzT@gID!EL2$%Ӧ{(gL pWkn\SDKIIKWi^9)N?[tLjV}}O͌:&c!JC{J` nKlȉW$)YLE%I:/8)*H|]}\E$V*#(G;3U-;q7KǰfξC?ke`~UK mtIC8^P߼fub8P銗KDi'U6K×5 .]H<$ ^D'!" b1D8,?tT q lKxDȜOY2S3ҁ%mo(YT\3}sѦoY=-- /IDd6Gs =[F۴'c,QAIٰ9JXOz);B= @%AIt0v[Ƿ&FJE͙A~IQ%iShnMІt.޿>q=$ts,cJZڗOx2c6 .1zҪR "^Q[ TF )㢥M-GicQ\BL(hO7zNa>>'(Kgc{>/MoD8q̒vv73'9pM&jV3=ɹvYƛ{3iψI4Kp5 d2oOgd||K>R1Qzi#f>夑3KմԔ萴%|xyr>ķx>{E>Z4Ӥ͋#+hI{hNZt 9`b˝`yB,Ȍ=6Z" 8L O)&On?7\7ix@ D_P"~GijbɠM&HtpR:4Si גt&ngb9%islԃ)Hc`ebw|Ī Zg_0FRYeO:F)O>UD;;MY,2ڨi"R"*R2s@AK/u5,b#u>cY^*xkJ7C~pۊ ~;ɰ@ՙ.rT?m0:;}d8ۈ ݨW>.[Vhi̒;̥_9$W!p.zu~9x۾vC;kN?WƟ+fx3SuKQqxST Ζ2%?T74a{N8;lr`$pZds=3jwlL Eڲ t|*n8[#yN SrA GYb8ZIaʼn8 #fg3i`F#5N 3q_M]j 8E!@1vցP7!|+R@;HspSI]ڻCZUcg5pDcIϹ,oN-_XI,3\j ]ٟ5~' SuipA!C厐$&k7dmhz/#"݃,YqCL$ڲ`"MUbeT>Xuv~4Le͢ }UVM)[A`b}mcE]LCEg=2ȴcmZ?E*-8nhױ1xR2ϫCya` A y!?h!9yL%VLU2gr26A!4vbSG ]ꧧWp/ &ee *w$-`J\ ptǣC^p#_`{ К8EW>*(D{ٛ,[fnY𱹞M=6&$<,"lX-Ǐ_whaE 98 (oѢ/Р΅ 7ցl6618ł_1/=fu).s¯?.S[{'g=Ҥ):d8h\y6]t1T7IUV:;.1& ,5΀j:<< +Y?58In'bXIǣO{&V\DŽ0,9f O_"[l:h¢8wݓ19\:f6:+ .3}=uvKc ٹeS<>ij(o'ciS<{1$E[nP b?8E'xv[K+E{,Qƙ1*dcs_Z'407|qBOgYU|U--sG8`u! qGYܷw;ȌCPc_|(RaIBKb+{P.T! =ĦiTob d<>SHr][KqWs7ѝBYǭ~RR"p9dFg|K- obY_vM 4>/]e/dy,8!xŋ5 R<^mYo 3c9(F?he:9[_v~\:P ؇'k01Q1jlX)/ΏL+NhBUx~Ga>Z"Q_wjTLRˀtL L+BT҂ll魳cf[L̎`;rK+S- (J[(6 b F? ZvƂcW+dˍ-m𢛲@ms~}3ɱ© R$ T5%:zZ甎܋)`ŰJ38!;NfHohVbK :S50exU}W`upHЍE_fNTU*q%bq@/5q0);F74~'*z[\M-~#aSmMÉB2Nnʇ)bAg`u2t"8U [tJYSk, "vu\h1Yhl~[mhm+F(g 6+YtHgd/}7m]Q!Mę5bR!JbV>&w6οH+NL$]p>8UU>Ѫg39Yg>OF9V?SAT~:gGt $*}aQ.Zi~%K\rfm$%ɪq(%W>*Hg>KStE)KS1z2"h%^NEN?  hxnd/)O{,:خcX1nIaJ/t4J\bƀWc-d4M^d/ ʂK0`v%"s#PCoT/*,:[4b=]N&, ,B82^WK9EHLPm))2.9ȱ  QAcBC-|$M\^B!`}M^t+C~Lb }D>{N{Vt)tpDN,FCz~$)*417l;V iэ(_,j]$9O+/Sh]ice wy\Mڗ$,DJ|lj*à␻,?XAe0bX@ h0[}BU0v']#Vo !ې: Z%ƶ(fl>'"Bg< 0^_d0Y@2!ӸfZ{Ibi/^cygwדzY'Ź$:fr;)ٔf ՠ3Kcxwg*EQU{$Sڸ3x~ 5clgSAW"X Pҿ.ظwyV}̒KX9U1>V..W%GX +Uvzg=npu{do#Vb4ra\sNC/T"*!k愨}plm@+@gSUX覽t01:)6kSL9Ug6rEr(3{ xRP8_S( $?uk| ]bP\vۗ晋cgLz2r~MMp!~~h?ljUc>rw}xxݸǻ*Wu{}M?\GSߋ2ꮺ5w"7U0)lۨB0ח*zW߬V}Z۫ܨJ<]B=\>V7¯8nq~q?A-?T_qOq?5-3 |q|w.dަ'/Y?> (<2y. ">8YAC| w&5fɹ(ȊVã50z)la.~LlQx[b&Pĥx BjIKn"@+z'}ũrDks^F\`%Di5~cZ*sXLqQ$q6v+jRcepO}[ s\VF5vROq%mX-RÈlб 6jf/AfN vRPػ.6<'"6dv .z{I>|&ׇ4Ăw4 [P{]"}r1殲)ߚA 2J1SGpw>ٕQѱ vb;pV ^WO+į1tq61W vzZ U'=҅}rZ:T#\_:ď);KX!LHuQ (6c94Ce|u$4a?"1] `Wa+m𢛲`Rs _I@U8jxɕͽf3[Pg%,IR Ř`QbmүcH&CLlvLҼé1ivGgJ+u7Τ!ljK1SpHR>:YF2cU(77eGG\ m#Tvmە8[,)4\\=V~?C~>_) cxF;;Ds'n [&8NJP5H2Զj{RC>he:ա+e/.I0\lWoӊĭYcxN^SPiMrFI_"*l§,̀+ å} .[c&SX( ( =X?D5ۙ@m cEpR?H0F>v6A*:W?*nzfw*B#d[se$U>tLNÔ+XX߇`cu0:U[tp^}{>H4z 4 (DtH-ʐ?sk7iIbΏ%T}v}e{aBs˞L=ilNeb]nltwfCEI"*S k`u ygz[~S [j3+sE.,uDΡ1R:Vݐ/CBc˾] shGՙf 2+);W{@dlG)%عF&4D&u.Im9c$A$Dfj-ء^6&#OȯTgرBӆI t[ 5)l>MR2ǂv JpU1cJpրj&*ߗEЍ0U#X) bpNVYSD1౱UR}UR,:lơ2<8"˓MlA2 KvP8 I7D Oj>;V|a|`U>D*KS;|:xI/ió21׭ȦS!e^t+28b$d:z4 .}gRcƈ^ʮC^0l[hl"য*6 ny!HQ=GOf"8vAq&*țTOWse~ (5TX%/8vS:w}[ą qf2Lυi lm/+QD4t.P*2V J`\g2%tJ4vX[7g"z{1|\*& >Vv:V^S7{{u%[^g=pn]Y#&ߓTί_z7e&ӃCx;xLh+NOEp";SB/eWٹ`64F 2AhF{Ɩ;>87DǍ-~e;\26Lة:*mUAN=VޮL> jwB}ѹ .MVfz0Ïd0l?7- }|>TT%9d-9UK=&l&~g&i"L{vrQۻou}q}hn+.{pWEqws]]|/ǫ\}/J.MLmc ԗWrU}/Ǜ+sYn[ﯾeywyY]]¨Kpx c./mo;ߟRy*4݀wm&8֨Or4 &+Bs=8'kP 3 |}44S8UXi;f;VE7e4AdX-fS烠1Uܦ$lznlq"җ^s RTn|RKm;ԻZ3)`S!9| ?}m*2@"G{yZ${˪A6yq>Elq*E< NX9@: Ih~|Y4sopp|v1f2춓tczF_/; ld2ĉwt&``4Mk"[뷊lٱӴI398bWvwOh$2т:N8KOzJd z l$MQbJW43?V5yLǪ H` QH#KywWؔ4ƪ(L\BIF/(usg*8;nOOyTH?!Ua*?o!sGJ أ`J4y NՌ/ bq|b ]] 1mCstivr]N/ u_6s}@KMaM+@yd*,E^)Ӊ90+)^LI *&2L1TK9 (I˰92k1WRoXeU=YX*^^)- FHu6bg1SjM-'KiՎ*D]4a QJIXg0K^ y̢jK.?$ArAL|U %MVϧiZW1=|;zB>,/(ly)w/@%OE]OZB;[Kxy4"=͝M9{>?=r0]dKJ<2t ĭ G2hDV<).[y6o$GmXG-K{RioL/'RNIs'dIXNZxu~zS cAyC!H(Lq\VU.v,k*j4,U/Dk{K՟Nlf͇9 7ПxʓƪPAsAVWYqy 3*V:Y>8_08[w+KCC[Xd'Mk6Ύ5 f.n:W^⦁A^}|Fo@e&8iOeXq`Q{.uQ)bgZKeg$DdXݫM\ ]xd8)q<9J_>2c_<ۇּ,e/ηU(^R4 S`+:<@qgO[H% y%vf<| lC12 a,?$?IZ[ ?k]mx< t x˭sc'muȰUyd{}z^Bχ3xjD;@!C4 W+] ,; Kh!'΃y:ykcbtgH#DuР au%Ϧ oQS.!(4`$(Rj IkqqԨAz_KqrG*rTG[1}#W<oxjIO}ݡ|wh(bG]4ntk?yh\6 b8sg=!?T z'|ѿm$1ߓ_޿zri\\ Gg)Aq%V%$n$ 7t,ݬ(ZhI"Uk!\ђxWGN )*v%e]E<)i"p }Z3+$"H=Hގ^%W|ӄW4w~KCʧx,22 b>>E*.qSw2+ /V3E,hPlS h=IV&gD4 mnq>;muh; +.p&Q2"jz-)HvRE'+p.}Өy*@mUHw(*5_Yh"G .szpv5OWX/YXS탠:4Jԝ}Y鰎鐋u\f )W" / ":`e#p})nݹj^UɆ&$5*A2 XPڒ=qo[? ~DHxn3ʊPj)MW65V/:rʝ1: r> Xw#!#^'1?~j`J \@5ѓbxR{Q￱DV=/: WT\X UO˸}\}' 33@c&RV=Ҁ0˝otKEylMɮEF*À #yaC!IWZ?7+k([7fv-% [JY y*-*?L16悩*]`2Ky)e]: Tsl\]7ҵU\5" /SljAazMϠw~q"[ K,TOx[NȂH2Q|X}|Et!ln›7ucTdzf_2^,j v%tӓp]ݡڽ^|[b$RUp۶\ IjX; {<Hӓ`Oj&ӵaNt,P* ꈆ5z YjnjKɨO Ed= Ә%jmʳ๹q2KM %@HB6WnFI].wc$+6 ,en_4;aNO%GnAmjʕtzRgKsSuo/W (jl+/nlr,U¸@$o .VWX$2=4dYiuiaH$ՐS%; ?4& gPI)uS$o{@Y3h}c/Ͷ " rz=cf0NR#XbCpwd7jSN HAK}9wr|c_[q=~>>dE,eehEh$=3-?:AUKv2Ď1F#j@c0ϙGSM[[EӺ"~gYkk{kKn:KrllW&gQ=ˏòB?bf]< = _w#F LS{M[Ct>ƂLW֛uzTlY%j[6 Cdo*iܘZ2U?Y9TTU p됖f):Dd7!m\͞p)EۇۗѝqΩ[DtF/tԒiL'dK$檈myM$ZٓXѓJrYnJ]}YbyH. 9ѷvW ʊ:pRZJZ^2-!-VFDዬH -U S0V6D"4e u QTm=Yܾ$"8>(ٓތ㣦|K`O}*]h`'.^A,ږwrsy٪%VU|V֌1y=$֤KlUUnNNQ)^qmU%0!y}l̗r%.g.Lk^аF ͷUE*)StmUHn6&2ej uu QA:Qeyd'ʀay0-G7eIyUk q& h.Wurʰ@}fxa5׶[6fք_,nO(>~vs-fQ@nfgm5snn<`$\@#;7|G2;;i,W{M2#<&vjv s{ M3Od`@`\689(싳G鵳=3P}-,$PfvژG=# vfH_P egD<2ݹv/dq[g/dPwog }:3A=_'!֡L,LY`~yxl P32ܝ&X-癘 nF5@Q&|&>l?9E̷,-D7 zNѳʤW/ :쁨2\ho'4}y]hm?EE3 t~Hur7|:xT0sɹCqBbYs,~M~典⠧4GN_Ov^mBHZ54`5&Cf~mu `+m@`;Mb , lv$QCJvFWER EZlOm:I{>0h+\65 !M*6Trӻ]q&*f4").|Y[q؊9@ eYn憀97<߿yW>aL^,v5 ”/-(gutI?=̒"Ve[uE;sz/4, }p,5@p3WS*ʲWEpR@?bn:'7n8t(}mІuwASMJ'^|];7i#UA-Թly?ώ%_t ;{˜r|{ "O׍.MȒXSEے0́r@ "vIPv{]r֒]堨2hv"ŦKpPP6cwnܥav +E}$Oi9~BHN4R ˏ@pus3Iq˒$+09ީ$h&9@OrmC_qI(NgV9 }UGVINwrH%a:HHB`{6&G$(EM@*$1X]MIbϨ @FIZa,Xags]az! II.s8ƾxH:Y,$n(86@9Ha5I(^1'R`L>?$",8rPTHa(zJB_;.]w3wyusbۻv=j6sk17my?Dn.UWnWۮضzkKn3hۦ}Vi[|ڞO.Hۢ;`[A@](]8:w(9ԗ4Cg9[D䎒;):mla_\AvyQ` /-33K$%)*79J.he%9Ԋ7' [^[l\oQa͓%]a/2g9sЋ npV}PGpz:/P:qqݞoY1-xHЅ [z=݌p#/JuIB 7>\ %$x} P,t2I0Q&'QA 0c'5-፱lư!j'67c6VoX`vևX*;FSWkZ;9Vk MdG]x?uLM*L65\8 ;BIb˵#6c E&Z&9>=Y~z-`#::;!cl{ ~~PI1.K&_dz[vvLA4+ggW Ğp_l7|07)r\X@;zMi`hĹjB(s'Q3q!N6="a|2K˕9Xuݭ;zHk6_P,@3xDT,%K^`waZǙ㾌nxrH0ͦ`0jb ^ WQ;K0 a3 lt[|P[Ǎ GX7~`^ `ue]8&FPZՅ T\jgm (l6Fھa?ml#cG-] {#? n74c!Ɠc'  7=XPO,Qfi\z+b+?N>݂0ß:cl6 )h %dP`k4S .m+jerFm إ]vmMopLx^~_I9UׄUBiBisB& )m2TbP3G+m4hdec+=MFXh@ќVy 5W 5j6'|$fBUBZ G%IՀP{PvsBjKj7 Y%i@ӜPy:: uW u6'}$nBUBz G'I׀PP~sBKo!tI/K }A*h4|VBPfT9BηRw=ͮAp&L]Zd9NKm&C*}t䎙QaܵHcJ_+zHTYZvZ}gnF?V\;t ᜼AfyZ$}:]|cKbQ[}4(>otr 1?O'!ƃ Qg׋y:fV\qr/ fy PQjC9E?$Ɵy4s02TvquTQ5Q` QDY?!"뭨N/J]]4MbKePɆ&ќ C`(";W#h>j|-vAnj|5~1Nd 5xG(SZ>ƣO-/_6?>^jL[V=Dcu)nGjn9~Ƹ H2,i|xm@{3ey8x;mŬn`׻ l}>m2IpC'ʴM"~`O<,n1ǣP׻,4TrIQ| p`c8Xx F.B/dģC0Ja vN/hNQ<i+3LW6O,f~o φlrfeN!> t&7L5+x_a xZ̧u)U.rj˝'AI{ѾIvaVX}oyƮ(kH-KtPy>/Va-l v{ӥy2;/t-}oA?zz@Fo61?heK椑{/z ˄9 G/IO FHj{O{f7ۚl5 /Wޜݖ𹌶gw&9K:QՏx Fb'mX/>k^h>@W,ݱ/<YT-e~:'LwV6I&c !w).ݱVbpVZЎpz0SN20c? 'L۬[ "B WLD,xETjpg01MQmqx$)A#w&ܒrf 3L4t4gtb1oZj`g I|Gq {l1F8 %,bJmX|qgF R_H^n儗,e }'_L5 eb78<×)Xs̖4jM?rNBiDci6lw)ޅC4,G/0'3&9S]>V%:/;4 ZuXniوoy-M9UR,"5ZJp\S!:Ȁ,؁(eSt£:) ^üaX처+ 3h!nyr.XYYmV%Z5My њ̛akK"o‡;Dgþ[=Yk{e,F֭5]K>0"\'p±0‘ Yk K~Oox+}$ v}>,'hk9R]LQkýo 䯽7ij4Kb~O:ݓhZgIݳ۟v)uA9?{z޸7')<7+lJKNk@9he/NZ(<8}SBd)Fϒ<xsp4Iu~d/zr4N-~^n()BF6(c-]!ڂ#~_HpCplN+[,D_bt6io8"$1ٱIL`q&,+#OI2d3#"Hp cڌVU$L2ddVpw5*[W$}LP[X~@шEL<Ā}g(ńUhV81X>w'XBm8JJp\ ,$bqjtX$v(3GgTfZpmjaDVXkw1(/v.aI3c2{iJcֹtyYSZ+v14T ~CVP^`:eɼ*-#=kϪzg=E=dk"s;gyqOI(lG H1c~~} YS$a֨tkg>eSKHƊIXLj5-k F\FX5Hbt'O'z1ö5ZcO)= n3ĦZ.[FH$o$ 0jZFV$I5idZ&}Hb&j%U+x =傷"}}}}7!Ufk2if}P`EKfDJ&V-M6Hg#, 3J"꫃Vjxtbd\I2`İ-:Ic^X!<*:".rX7E jQ\ל?h8-a_/fQPZ2k3 bR "1lpzYpeB'qHJ.Ǟ0nv|*#V"@}A muC.9v.Z_TׇE325ƫ|$7M> r(qR |!T g)D8W#k>{҄Eί7S *8z=|[K,BLgm}Je1wqƅ;d|V(x5bF,]Oɰv7r:eY zR]"՝)w8@ rsw453&v.fSbXD1h A4-܆XJ\ DJńf1zMzNFGRs5)njދ,FT 1̄cMuXT'l0s$8nّH;] cs hcdv,x´,"9S\!jdG6B|p8B&cLs[?Iʫ'Ό^5f,i{}i `$ZYPe?U9[I S=!U!/2D%6 ׎*29[c@⇑D95\EO 6P,VkLU& OfHpݺڜVZ_p}tKLSIr]\<waX\'{v{ ;E[6pY5 HOlY#݄`vaĐ6Hg͢\w||A Wyd +xAC{0$!YBH1^IpJJce~BHg[ IHzMRҘ?g\tJdjS`A]/Jy%k}JDJb~P+f 2#qދ -?x]D3y= )OZ5xϢ̂Ŧ"i4)Ĉa>$eͰ?^dFcV"^饑,X}B$#n kdE }lCvcW{WLvk&Yf}g̮}$8e m9E;83"Ě\^XpdL5H0.A5Na_I>s1q,|Gp;{_wSYiTq}B-Њۛ5V[0QTZZ#9}gWϣ;D^o"E~.xmCl L<}cxC>_eyv[gC/Y1)hX})&H@ czM`R]x%lS qi X^Ʋ ue?L:^ɠvF1m^IU{WyTbb5zb_ F/ 6:s d`2NaR¾=rk+efԁr9y_v{3;n4kYR ~_HT%}%!Dj=VrYlwG `)o+vn͵8L-kyGSGS7hxpH2Bk@ 5~̎quQ&b1pGIl f(w%exw5#KԃKTnz]Y5Ihfk0v$8.S 픵&40~"}s3)[{+gBl0.燗#11|@1 5x}FqDbW*3xAޯ j+^ )]{F'vZFg@,M -fܔQyQ ܙҴFfi0.Wٓ8:KsY z#o{Az|*+{-n[ nm<ޓXa+ț|au95NЕO^Zk yX%c^,<1c!pt<4du|kfU>hF5$8 VUy 4)!j0ȻDA^5pmgL1 uneM[ ô\ʃ!qOs0a|/w$b֯ \klnދ@syeSVU 0M11~!aLeXTmn;S;ShNH<N-lw#)p .jw;q et[F^hps`Gp$! k-gdϛI.]@N]нijie .a0Fe݅u"QE|,ФjiE~bMZ]1j<>/n(}FA5fwken6c_b]$W)p ײXߟF52@j;$ d1 L!k1UL6KeV8v́6fl?f:/Ip*3%EbduJL]0ݠ8?ZpTd#4o$8o1K\+{f;LEݝfHpԃʸebM*,Ig^T!t^JFB?⸹{~ޑhOޭHF9 tfjj4.Q2KN-5Ht؏`hHj38׻jtVjY%󻈪' εUѷ1(uqok^'gHCJb^ḼEh|$B^$uZ v["7Yj=1z>&\XBB҅ѽ Hc60&;hd-䢚gLcc6zp焔[֘u0t‡:MIiq>S1|,`3[!ͪ;5e  vYtY[_//$8ހ72]0:%-3-j=7#IjH×ͅ$?*qn)D#z=jwdINШC1,VL c[EtYP$<1aq#Y+KF ^v`d2ha1"&)d mHu] Iڻz]/{٬Gg"ԍJMPv`KZ01_JI&vؿ{Ҫ/!!+CLU/yp=B 4< pIdb!!NV)#8VF. ?N`_\NK\+o9|_Lm-qe:+? ϲ#tZ5؛ee£|{i6"Y2{{ Ғ?@Μ;(ѸNyH5}!~M2 c࿀ih^:e,X '^FHs MrUNV+ 8{*POXx^ĜW1rIb^tQ "d4h8a3#x{[zUV-kahbY9MjXO-Q*K$%8 .Sy 8 \pxA\yg?sh8m htg 8qYOkσO$?1ƴuozk [")480` ӆx-bPv5<~4> &y3@$ h9ÿw'q$F@GUvc4^Dx\ ؎# 8td :8I&&}18oC<17Q!0+0Okrɻh8*dl/fbw-/v E ;0phNph=]1`VEeQPQUrUqhQp=Y?^]tb Ɠ\sʪwYfy~]hV_ECr8319 ̛\̊# ̎Mzc3Q8|6^wR1U]N{UAq le%s#L{J-ͮ| CNKUT J,1q,‚;̟?7JM7q1T# Gy'W~)Fz}y1g;05^h ꊵEe . IoV^_r>Pze2Aj9#b5?Mt>KJ9@K59)h,+F (^aׁE..wz;ݿQnQ9sxWON@JėS.JO_M.\Y|<'YG֒[D0N:8uhwywZ%F㪖c*,K|6Q9W)(p猔(B{2{-twE?P?{\n%-oǺK%MК^a}L+R2hݙfk jAlkVth}S7⩕8ߵ\Kgls`) +ý:,!(Rff!3ZA}U;^AR`;lM>+—7+VZAcMx%Ǹm䣖slfFhbKʼn4p2 ?A9lʉ tLD 7>'DWb1(4MW Fd{=0qՍձ )Ks"0iL`"aSfR mi'gL+GNu;5Vv^28spXZqq/ N2b)@hd=%| 3PJUŁ5{@"TlUO&=E)0)ۋ$L}'@Mpm>0I8"DwA!=œa&-=3ƿ37MN9B&$ywH@bP \D(M!"m3_ztox|~:@Jb (X;O A1Uo{f0\^C(ifR;3ߟ2dOA:HܻW] Ŏq?I/$LDL9?fCmZvufiZҧlh1Wg蚊HrLIK49҄y|, ?D#}oXQ a84D`6[xD~"6+[&XL+V]YEe} s+/w9pT6l9e1'i~t9p&+6_l4?eI!-"tu{x !.A+ Z#SB X!.w9(KSmgT zn319jjYV]ؿjD* L=D_7`t؀xiyU:ܶLa۔PY #AX `gy ڊXKoA%mܟĶcWjZW8nBb-ZEro$٦k` [4uzjN0RdvVIX* Hp"k &@QljHJP>t+nm-ږ_jixO#[3f#,_[Nʱ LsA*Vj:Υygy0e \6ALމSuiHQ`4E^$(MP)D4`CNSzȲԻfn Qe$W!& \[J9AJxR`AU!XÈS8%FAU>;@c/0&Ɗǚ|@zip&%T&!E59~v+ח٘aI6E[(S&rJ[ᛰ)K2ۉӴ]5K(!jA#0Q;n̗$z`]yF+A&ud2lMiYN:a3[]s %wb},t%~:Y[8|/H?}̓z@0zŤKT 09 BI.R |f%Wo6bICI4;|LЁf!nRLAR=]-R*!pa $ADj$HZ+\ƶSɔ0v̤(@,[@BS* ZQr5aXƜZ+ 5kRSLZao1YXLמ\!Nƃ~G{@QxTek I-U-&)=QEP΍myT3~BGT)ƌLRP&$K,X44MN&FHE*Ewzj-@ә;-,ȜLזd"61i$ȦQiMʔXIƱ4?2>R%Di 'u)M$db RBa<+\ W,r*b NECR"Aǣz"MJG!I]#Ȥ(miK$Qbaׁi^/ܶc6]FOGH{ l* &d,[iJޠf),+rfբlxuFd-:0j6\INEp6qBtL!7$F23VqbQc6,x8{QToYo7m{/i "9A:JiEKZ>!-̩UhLE(e JbDSʔDT7g\:0Zl[oԃ[ZP}oiN - TUpdܶoj䄫O6 "OE-3ǟh5"ы2qBbh!tᲓ1PB^0'/T$(Wh1tcHK$0lsc([VWA\M@%m^ ee˕52ZfA yKB*07=(ind0pw<5~-tM! l:MohB-AݣL( )5w^(?#>}8_+-Y}H-R7ɟx 8vGY77Ô))&sO;oxgb=~=tj_Twn97. `iv(΅u<)5WQ~*:^P4I[Z &̜d`EH¥T$K+<zm[ ԂiP'$mX6kZc"ZYz|PR&\rUD Yb,CӂpnϑO+xf`5tXlb[t$DVl!,'G:DFT8@BX&?gjVX\*OF!T/ Xw2ѓBU|`ʵYi.ut<*_ 0M(kk]؛.|ok(]j^yMrA]ZG/ }{ ilǹ>\ ;l-* yۇck/_ѿFC! a/O.;1Ζ٫ zOAK :(9qAd:^@M0wQ#i?LAuҫIX}Pw?O+ ЁTz҈)M2xՋWϫ].OX CB)W<ƣ<*TGW'@Ams?L;!ub1*f,g(k=*>n.:-qX>?T|d@g利hro[\C@p}NzNA3#^9xyVǜ0):i}Tv6kԂ` GX?[m9[``1>AFn)jYPWl/$u\{x.SLOifkjmFΠ_{ bkF2,3%B@ Rx(vfiiF3i]|FU[0[veY=bP \#d@9\{q-|ŝ⸥ǻKyxTYfWjqS&QGٔ_x3B򖏚IJ6LK]|&p&H6.`I[>j6&ZI>|axY9 Zǜg00\H`|yTڱϞ538fuih",sB{  y2 Ju*֍c9'\ 3`6zܧ[h%&EkՁXj5^K7rA E9#| p XTTϓ(/hvǞIySJ{۟t [| vTOO[*' lJrP rPL_(*4yX4*ji׷jjOWǬ! &wP<P^|:%7LPf:^(+_ qhR)u/{3 WD$="6e3'M T9KUAofCAflQŢ]}FQƛ(_A)MGyE-y/OO_wO" Y<`џC z9  l>|YBd)n$au F] Eʠ0U4y"JXc#8ܕoC7hp3h4}roT0Zލ Qʣys6qdtҳe z$x>s,m#ZKcGiAuLkW JSY_Ί9}Ō)/̯g^4XhpuU6̑kbfĝ u6070X MquXݗ'Q`Ow L&~4-mhD`iQП:o96!Y4l֋;*/Ud^ϓ>.AtWMQY>pn {P}M)O]s$(:D,eiS>&CMxD$7I!3驜bh|/n_79T X'+n-boO*8!r0ųn↉a"Z,חop³7³pK7kÍ_sSᠣ >o = >]GCx L)Wpj_*>9zKa iٳ(Pa7g.9w޺I!ُiޓqh-h1/?{ q;f 4T ȓt\'*,d ~0.8 }n&%;_XԃfDn^|:)n- Ze*;xOj`(/;g}]ibLۤ+4uLnm[`vv!m".6&GCI$_)>%Dt yp/ܴ%,U{{zs&srg3#~>?[<(3U@|K^ Edt%7|&h"_[j}:gЧE ix6MyK̤𙗞CW[xB $L榵_XZ|$Aj6ȕ{<'5~a tbkUy2#{y/.b0 |4 EU-WehQuEBl7fL8,4Q֔V}k-ymMFQnKQӻ+ETw@ԟN4cP"u'ij) xtH6b^MY̻4X]ZαصZ)6#]yU]Yk,MM CoW.̄Zq:ft[D"2P;9/LFi7zddUpPd5/3Ἅȯ1%Lkr BmJ5uDzv_j ~.K-eUɣw5׈m,pl~fӆƨdTQX bSr*”/[六 _丘=AO!wL_ҡsqvx vⶉh[} @QAMG `IvuЂi~ݲe).O_nۡ{<-%i҇F!jq[.c fCғ9M~o0 Y |ĥ60)6HgўV)> eo_FCW}|iE9aΛQ8;YÇeI;?"ȘDvcgn=tS};ßŝWwM_ 8v4L__5_R@|7;5mj)wE,vWO2d% Gq@X7'Xz݂ͪ_߿M>fY֦<2͛uO! 槏g $[{7M]HO&ZrDz̕U+q8Q2^ixˈI:G7҄85XGhD3; .YDT4IfTjeLuXx9gv#9,xs1>xmc]'k<ԐMIod_>jbR4@"##""Q>|~Wyq[6g0akw}&Zxpssv;r;1xp֑0&nZz ડ `gW9|?=k[ٛwO0Bq5qumNudc}o.#!ӷ?{/v'TO%=5[p(W[++7cS]/#+/~Xoc*&+%RD(pљI18:,ǁd8Ȭz[oN逳 JsH I!N {\9T(dt>ɘ po7ނoTG DxDhŴ4p1k;Ł寚N90>Ge,F'lޞe<̓l!Rq LYAp]Hvޚk[\+ ADFH0 .6cY;r}fqc,4ϠB+f!XxF@ī`H51'=4v,>\ַx{ʸ-“@եW*‰0 =4v,>zBcہgmFॵ6d=qE"w! i`&cX=(cGrװI,%dDڸASDرrXu1?OIYncIuo/顱c9G8,l:ѣ$tWro_ƒH"]%0efb],8@"l`g \6DE r+7P#8mSjJcWG;1qН߲j:YzllTNdR;4[GG޸F,(qLVtWx^k#@W3K5ǣ|B؛]>\V{EՊݙl{btS_NJYT5'%uEI< /iMd.q%I뎳rvB75Wˎ|{I׋oBTTHIBwy0 `"D옎Aa:|,rA/?(d'14 e$!JFJJJH^#pt8 |$˹.XZQ ( I<'#p0Qq=wz1a҆CY EC^qhvာ|:o]+sNr\ +\@J`ʂ pOٌXe3 |,դ]u o{̫Jk@0X-4fM EOHM#q:QB%q8@0>݅U<6!H5%)xDBT,\;ͷ4!G0Dec4Y2{hl1 5铀?YN"%E,s`I'!<K{QDz\Og/֞׫!v9JAmY5QsV}r6VkT"QVĬɴjTiL.!Ӟ ݓHYzP, ]v.g4Y.ɡD" Y'.Λ]wqck)YyQוT|?;`滿d]|[%CQ9p}9ص-sQkq_};?n2m0uyQ+hF{wlo }C/ܞoyms{r^ Jq 8i$ZHV/WLcPjmX6˹-ݧ-Q4i$c4 {Bf3op Voiɖ^C{(,gl0b"!Hx86עlAp|n?a@FzXӫ> tiS?:^)&Fc-K 5μyQTiMys'.6 ?O=4vhC!"nJC/Wm?< K;HvUqe(ƎGR{A_ 龈#pBR6CMz Jx\ŀ.R6V"RҀpɜA4W!G*DescW#dz<$.W̉_tl\aF%=4lJ6H;.a*ӈuu\V+JtuHPLjUJS6K4Ⱥ4Q#wn>ntCaᦩ0D|B-#j4*;t}:DN,}EoT6oLO?3bp"N2:= .Zz-Lu pGd`S}xPWwhBm|ehنcLSu 9ɯ>0Ci=-D;t!H ^ 0ve |^(:a r\3(AIuP C&NxGKtwBEPQcY+dSsPQz5(}RՀ#"Ѳ!'i 8TA0glh1]w<[xk߲J'읙珤raq(rea&%}&`D _;6`̄9$зQà+ ΀2^ήI> 5R{hp N:1!.)$ C4vžX@CIH#1N!AHX CYq<;Ga@Fa+ h~ƁԈRbOC^q0 >BòJbZVV\n!$[W&)=顱ADz|º%izh4,i,e;2ɫ=~Q ;- j@TC,p(%_RQ7my.ov{MXA\R! Ku@u9GX'mUbm |,_{)hn)N@%49/_(}=KTl Hrv1%t?cU2Xeˊxp"ʔN- Be\ժ|Ze5"l3 %סgcҚjn |,'*i)߸kMJ,+D3 hnD N/Xa k ~y΁H9p1mTHge*3|,'\aUCRt+9dx%SwýTp m Xm-p¥Y>_^}kƀea@SߢF eڙЌ bDO~8_e\\߯uЭcC#в',E*h,h F>^kk&M2[_k 4>|ЫF]p(Oz|WZ׽Ր(L9 '1Ɔ77UAn" ɱw>KکbGrwʡ`7%I>5<ױn͘3g̙[1[gԓՆP:xn=0*5ݽF˓ޭ`m-Lfd;qCy?q7q rКv?&"VGk#zzz=Eco㧘"S@o¦Tn3 lKKqKgw<ג߱e(!?~ޑٓ5&\wcMU__͊ 9ϫu^ nWw}3I+vn1_Sg_~;_s9) 7p *=lX_x1S.a_K`Sj,+)'b _ Ƀ﹕{Nfn!fM~̋ټ~+/4 @U ϮU\/WW^|t9F))ao!W7>fIT>" X-vAdB!\X0> K  8e1s47Ai:b 8wZ382DJaI$agb)cĢn>rՀdR^x)\egOl̼b}:-8rH |ND4(}tiNaaQL:sFgm9˚`c)$ցxL)_\!,*,(ܝPZ<7ü~FIx^[ W gzǍ_1 &)R$X->veERv[==3A%[Vے]nd/K,,ů 5^{F#^ܤd!IL])%o΋A<̋A_*d 3}z@;3(IM67d QJtj$Ia,:b&x9Q h3UE~tEA pUAJIX+(hNjl(Ɨ@{j`?: ཎ@ DIQD՗lQlXy2WN>|ȇQ 4YQzb|l^vQkcFUG2>ܐSi1mG僂=( ǵ DU1b r1>[UQ1`\Ѡh4c }:B1^;<>Ws?/ss6>3gh^ @9a2Cw'^ o7ǂqYA;=gp񞰰 |/bFlxWjθع  ? 7KN熰$ٻM|}7 {Zn6/'fò"Zf"HCn}znVoPupw[~/aoOyZZF{sԐLf\ UYk֡Yxa(S%_K2 \?)?kZW\~ k]$e]?8%~m|u/?Ov~ގ/~֡l)=FF^&On8_&6䯝ou'b"^~R/u_ 1UꗟBsؔuh'w`f@|tpKI'gU/c}gL?ؐ %.[`[JSx!a!Ɍ>s&Ex TzkJr_Oٙ→Su;vӼ(ks 0,}EU>O5Lol֪ן¤8%W&Eܨ]bd)" ~%+~x7FQYJV79\!!#hݺǰ?%M{0z?[M׫Oms$dzG}&Ϟ|V}zܾ{\Ÿ?Ǐk͞wp3^3f<.m3a ~!]~n pE6+|е O(zs;=0X8*[4=gg3hj Oʾ-L=_`yb;!mźbpsøzO.kr{F}X|:6an9yN!NypEG\f+&41)R0FpvQ[CsgL fhE~P{ܸo\mn1dӧ9 c̆:N_6aJ, F 5ͭ$ KnIO0O6d#TfQHs<Ƌ7S؁j l`q!|kxU iG=q," ,#F`%܂Er 5|Z}4>5s%ZR`NhIV-%ZocV䔾"%YoXb(S0BU  G,u``9i6\)Jj֋٧5ߞ!KnU=~ii"%p4UxWO M Ta c.)j]sPPǛ4m K5md=;iJZp]fU g:lx鲈Ikbw 1˻0-ez9B8送{[3:3fXFv%7 &"@R = Fn >dV=Tبބ|=5ՃiU={׳ԛO}:9ʛt(obg䶡֦ocg ЉWO6vxRtrq X#̝$YЖd`t:RU8"mM\/Skm;0k"vryPyMsXDPyRwxk-/6{矨tz&59}:*"J0`6j.ĬM'>:*j81A$IXNw\E:xBV@'oCA>{ÛyFGm9`o9^%1IdQf.H'"}1-#.$<ڵ^<\5#ҲYvi/TXtכWmT|M;>MJH߇ji:>u{WVebx=y a.dDd,(޹L dDӁ~ywpzψ%LYF>$'VfS7NԆGy *~FRrr()%(Ͻ IDSL*'F{~S9o^ ~eFPznY?(A B6n?]E.xb2'eZ|n2]O¼.{?v~|DPr~»IQByoiv=QTo:>@вĤ t!/V_.qUV.堪ÿ@UT?O {ksTd^jeIˌ)>N\Ȍ)t/xׁWvocu88jkk駻ۖ=56B;m ߴ)Ovǹu0B7DQ~vY]38w$a* nJ7xULZW9SU1"TmG0hٿXom#]3p`-<7UOq… d2e!yT30Cwi6]U 3؈jۧhãKs8Z? ?s$יo(ۓšPP}j> \;F~Q]jl1RymFyާ ?uQӣҍ.1j]t҇kLME)OM 3ARm`,6.<`x@O͊zN Z(bԬqh8_w5/4_Nm>7[{1 pI:>:ȷvqgG1GѶ~ ?<>d+;D˙-9sk[Ɉ .qw';"ߙ:ԖU [IL뵺PJd+,hи^ r\z4%)SJw]U=umu(vu6BދІAQfλ~i>~{Ϩw!Y72S&Hf\fp l[YHm;3AݒZZR֐`@Hd5/UŪwx(dYc*qU[ֻ4R0- J $<9AH(e4 MΈɃ*ȡJZߗ9$ͼa WM:cS&3g!Iy82ҝw9IjYFÁb|4| duPV2nAKB 9N;y*&=^)#?+'xUJ˕\Ĕ"=|hI,*1Xb2p f2yt> 9v](>hINGBDM1'(Npf=sCt Gk$p-Lt_IՕÔ,Yq̶tf$Ty@HBgّЃ2u060cz]tyy8jCM+jσ4Ҿ*} %? VXɂİ8b |6+wfllEC 9Vh \a#R%CJv2z;}0hk1|] Nu Rsv8tEW|vȱ %IVjfNŧyᔺOLc!JݗȡJ偈% t:ZN}-@ldcH~)(5ArtsTT(׋q8Mx2 vw+:F$p(hK^moldYh,Mq*y, NFrŎ ʅAcDFM 㲰Jd<[ƒT r! Ն<-f@c_NWJvnMq6P5aP :ISS0\JSQhȘƕ5H c3̊0p t?![F C2?ilxp&  Rb *eeRNۮQpFF& e1NQGxU:y6c?Y$*cɅxG[]ja t<7凘ϲ02f%CM~ a*S#F҂}cQ vMdӅ2z ck=l2JUrsq>8: bVsURc9'ofl>,T;@]w|JdG/`#R˻IAuJܱBI/JQt"ʅwڊ`:n>BC$)& Ag<Mff>gP1NDE+ux,M22ǚܹǪn׸q:81lȲEJ /5\ ur5,-Oup&b9{(+75 I”X,>5GVc`l|QP!%F~f^DT 1@݅p.v9Ovz i3)]gv=0F9OV>R$Ǯ';]YA"`r. :Y!"d#n +x ]KN/ץCH 9tN&A*g hK\Eg#D;*b Pe$J:0ePӂ a|v9Ur'W[|`:lY.Mdpi O(Ru*8.*UMR蔡SHԽ4>-u~6r)&CTO UTLgX`32 L2x~A 49;f3P0zHHeWH*Spi* Wr[K nd"}PAdRhS>PpN~F(X A?VC͹)0<᳠g3 * }Α]$(v6_ OHP>RjwAǔ"2amk*VARH"f$- )q,Sd&d<,^u.B|pqciS%L$:??!7^\'ÖC-dcؕ&(!}ҰLT|dvOX] GDŽKˏ7V as!&E0DJl)ۏXS=-:_w!-9cMiDžeJmN] SK[ vY|B<8If)[09@B xv3;Yf r~Q`A@_YbЯr+y'LK*q,hFWA[ I9cQF{,HYZb/;CTV?[1 O#;Or)c #@Z"#E@HN-Y66hK+\AЅ۲cJR]P:Ј@[1zH4%nν'h.97CaW`#Et.KBWOH6nGv@CR̋bDC{ H[~6lL!H|yLi2hڬ1\ ) t:iv^/l>z U%o :F {,KgKvܱbٲc,d!)/ lM. Kx»*Pvgs|[ݧLBStrw7!Uvȱs[ ?3!+S9\T鍢#tPΝ+^+Ds*Ja_n'@Z}!ô$t`{s{@XzoU%T0ZZl%$Va6)o,<1&4nS /ER;@q!l-Y6%S0^T-;:p~⎷89k/O-uxPۋtanY< C+J.10P*8*lb{iKV4u&ldK1?%+ޏ%П|y4m@oOq M|+w_ ~5ߎEU\t~vȅ>73z (vG߯jfomfA%?& kdը &-yؔBN6q5AP~5xݛ?fa`/GqhVx2J)F@Uft_(yCd@}O#2Lof =[{lW=8sG}ɞBfKpr qu4e&yrJ|NuB|lP0Lz8|k4-u~6rXD̙)ba{"iX)41lvă7ҢEX ċAB [YLQ)Lȋbv[%x y-Jr:_ft䠵Syaɳ LvxK `Awf+ưVjǰn{UݬW(+k`.f_w:< n%T k|A[寜֩ Ah6I?L&}U'f%J`NY骰"[P&R*^ 8nbYʘ%ahi#uy}\_*5 .*>Y&t*\CP ]msyo߷)r%H?e;ç9x^wq3 B@pRKDUs-Rie-`/k)Tm QO^'0tgkS'(g#vT!ô\|7 QS/Br#x`KaNnF;<ƊN_֦~X߽[ [g3_6B #w6/Va#bU͙_wOAWզ˻uG~{]LĿooƮG5^FXW)VC-l6&;~*~-w-N})_8{7.Ԇ熪4q~͢f6MxŲe說A3I~*{y߼q M̲a1ma0|LCG{}K73Y͏gg݇M , qӸ=LqVfsJj>Oqvԅ^=rzV#QX2Z* 2'%g4W(뼩+4TSKKj8FL9tz> 3u(|oF:5[n⦊Cnٮ(4?P/n Siz}qH߄֔ jGcmQX~0ݦk_dWޔZ :4:֏h׋J[^jMyP5#}@n[ S-I`շb9~?n?:^FSpu/5Uh5!RkCF͍ZF]YOlJ*oHQP*JLXU%x٬S;naCC4}Cz7lj#Ϭ s[nf wWk<&RǃV%%XQS3yhyV_xw/d&h#5/MfPN׏$ ]pv: elmTsқF1 FU8{ݫl&22{87qb4 N!+(0W%VY+B),-_z !&7ldhjЂ4(A71.ulũB[8!B%`7;E4柎nۦRm򓌧;FI0d֤*h=xFCt'V9$/dW(+[`ݧN/FGL.+ -aEXWHb9ެ¡$J0=kv7d&eQDBVX^_t1ng%CɧuJdBYKA1@ "]bP+g½$Anw !d ; +|{>ZRg#DZUJV3h0L0Ƅ. å+tbn h 7@ΓN,d`̃h>c[Yr `'Y{@MX8TCRg#)HJ &2D]1D(aq:OÄ؅QBa IA@BEҬpR>yO<ph}iaKI27|1 8v:[r "\̛8sWWO^o@JKhkW >E_|~ً߷{?7 a&]/]aIdq|69wExU!$V7Gb>nF?_j|uq <|+)EU!)҅T1kʍ@[\@E ATPP.eAP.z_8%ur=As,=ίT6 \O9̱̑?c I.&TO1L)BJ̭SkrR1 N_Is%/+%Sc=;yx4N:d:еTͳx:^%4]qH֏&2v4zQadޕ6kR]5av/Mv/hpU%$m =lҵy%ʭE7.ߏwo۷sJ&í_?'wv?gw$ A.~B7q3-d Z9/ _?GylEl]pT 9VRpcג ΅I@AoAFFPDw+H%9WkH'a6u- r͛HH/nշ VuۮF}=P;Pe}^Mc6)P8S\Η`: G{YTƗ \(!p$X( FԦ\E"x d9 >lX?#ڀq?׾պmMW8\v讳 1 ׫}᫽yC~q%W؛?񌕫*)WS)OS +@LǏŰ`pf0Q"'Z=u^Ą=Eh _(o; ˩4Ts/VM&Jo_~̯ŪLC~ts>5~+ށ6x~z?UZ:?4Lo2I'ޖU}.J&d}݄ͼ ([nplM> >k7d+q|dqIt{tz[6bhsvS=ymSъJޭeK` i5-n%!uÎQ-*ۢ.'Y@i[.4c^'9o ]6CWݫUW|K& SN9Wu&2sGƟ)v qisl۽nej8>?sFNoߦU'X>#fgpp(7oe. F0,HAOyoY&EI7 wi~o'saP4FȀ=MJ\H4ƉsV%1w+?2E 7AKAD}qX$7.Kӻ8)<"ELxq5r^a+%7恫Dt|ƸZ皬!0 S1QJ˅sd'@۸VY-(w7(%S`X^=Fp)$Jr"YeuY+,D˛ #}(1J ?G s61y 8b㴲F7DlBLa\4YǥH]1*ڜL&~,0 #E/b؇9ɀʐ\p4 tJH[f'M f=Lau׭"%>*Uuh U}Y2S{ ;7O1p#QdY+6OŔDF+1ڤ5Fwk7CLbl-6H%Cޝc_tYfī 0tHA) [1 "Nk%Z6m Td%Ys} `@p;,2cWhy;d4zoޣ"秢_LCH]jZZHS?x#GĐKz劣zTܮ~XrN;VYzP_At-aF nX%v*J5t iKTi發n q4!=Ĝ{ Xld4B3ꈊJ#bUP;呑S_ (<B ݟ 5Ex5|*H$3tApnޚE"4l WOI|-1%nRg;<$_d^z\th ȕO=:}8;d4ǐ1O5{$(^ }0`d>E,eZ #E Cf$5 ]OMР8Thu$8>䁷}XO dj|"Y!zY,(GP%83R- r`e8ch̆6V9}W=Zg!185G#XRs> ixx@HVč0~> c(oKLEc;d4G\7C6!Z]![^)MdqzL*t8EN.FJ$]3ڸ:0'攩EeCM̄y7O=E}{:d4:g$Km2#&r`YщǥXq-.%fq VK +iPq[wG{J Y+Nާ)AE"9Bf#1x6XLR[j|C9:d4^dYQΐL]xU]i] {?)iE=Lnt`v ΔBx6^fw%{ CP(Iy#iuh Nw[\~mCcKq61g8H\[L` zH3K @:eQ]. Z?[c'wh Nl3\'\(xܶ@ DͧYq'(QnлQF3[._sPbs_SR"!)8'}c44 s5Oc]JZ+nG0=$у!hP4{N($bxR`52pB2SC=nf!yQmR-47yȚd:Ppz4 uh sݳ!x=ʽ52_=ꏊB'2e"s]WtXeړ9Vo :nZF#rl3\{ R [1"!آFڔ40cVHQ(4>#<'FyFFcpx4ez= J&vh N ުdV.]1 _B Jh gL>AMRd]N$R%.$NIL`1ky[S+4Sі;NHDN\`^cRݤ~]zQth N썗g=ՔJxvc0xuX0`ƀMӏEJTշ9ol+]# {{ ?͖h.14ZN䊚 Fѵzoq!v`'՟#UֺCFcp(-3 `ޒaT%2G..N\Ua/_~>tB 3[#-.- [tIq]brDQ&GPqQ4LdM^=B3F"A9r-`ֆtTu)]V?xG_Ȓ nTh}x+4ʬustM$d $SjO\smּq6< <-Cg X 5I-6^: _"KVkLq=c#8Mcw"{J?_>̂]n|vfnٛto=Coy {;K6CڗD۷:l9ɕuKX`gU̧H(vmwQ-Eg^v,'ibJ~Ȏd;< &d<RYjSDUy>yҿ~m@頱Xʯ}4᧵"[ﬔe#]?E}qx@-.y}f鹥iԫߺ ϝW0S TYqsz]ρV:7y,~J{lRbl3N*<]=Ge6)nph#r<%1}$g .Og~z=E3Uoa*hX\ղo\ܟ|)6w`:gI>ʏȽ.`Y;VZh^~Wkh|?/tF}SE[+^wX.z/KpAUkxxsNeK`Pgdžwz:ގ&lrbQқ<_e{Oybx^fX s<1`~\򣢶n,6'o30E-m1`՗-;nY>_TOJzmCr8{8[^YʢG&`^~ TvFEl'3;I_=22 Znn џnpSP:M0jdzMҖ>`\,q^&k uA{AO?Lg >>BKY8m}u,BB .b$o`ͬsLv1؋I0x(.$PB$h%=9,i '{uf^`-])Zm>nק8+Op yK,ݯfU)ŇU:%^TTR40N =rfǒ_EEb_-Mx}2Vr`AʥaV,8pKRNڕvwW$`P  (]/SrsL(RCs5 g1n `%#"2h ]ĀGIbOpV ?SX.E\`݋u/"ֽXu $x\ȬbO’Sd58+] a,A "!dS7N~ 8G:O-̦2jvtT6.a,nA빹v&Ygi_Wjgp"F"_Qѯs 5i(̜!)0UvvR JLt !d3vtwu!#Gօ'z3`#:<֦*Dx5R/EE.bxE=.IڟH4B1HD^Z40)?~%pE5ߪ{|@6/7P}]jݿIdlb/e*}ΧAt+O4Gs:J`'6\}~((Bf N-:"kw"I_yΦ@琁d,;@?bfUTA#GeHq\%[}>,/?.]gI Svd.H,C:fFŧCڑjAƎKqdi;<^z4OCŲcBWh&>h&>hDA}{ʇ#0F͂bQاzJD j =oDIÑ_O\Ȯ.쇒6VszoGY/M9($m˨,7* 21 q&)?y'w7}1zvx>lO*]dqX9͡b~u}yF^nzyrsĒM) q\pɀbsͭolof"x;8M^.O!%%A2{só%E/,XPX(VCLsj)3N(PZ g0d 섓w7tu `b6z`MAdhc*ox߸ m"|4nzQ,z[0&{[eNY!*P^ze+$\q+RJE iQbIj!sIox3pC..X^c,'Bp,&L VIgG ,#;%UTy, ҏ"E؁WXa*l)K5x=A 1bHpp|z,8۳K0=4#HьH"8Vf,0Cc PLd3[ .P0Lԭ eغ"<6JCXR,%U$(c&$H KZnΛn4*+2;˘Ĩ]^2 _4q3>i1|#c0f"ra2ǖ3?$=7|L_˝ Oj P~K>,̮'kR,㪙˧=<?Cf.66C^L\NIK._+[ul^+lsWmnA>;ԥ%eꃟzɟ|Sh1O1yV)Wxp7[Šu,fl*[tP9Pu[yzddP>lH4Ev%8{ ڃ5Bl.3]g U:)355Vk7#.,7+IȡE~4\ځU %#"?Oti(7ujWcϷ1,/"~i#FUgIBvFQ빬Yή-1y}stR_ 0l371+!|9,L=@bRujW 6~;O~ 4ؙ0j32zKfg+h[gQ"7}-^;:E&^LclHRM Jg̴S qH3=r; @6i)>^t+`"9,b[|h#89)yYýNLid-`)d+X~޾#~T%HrOyyTÕ.jGy_0ay191M1!%TH-ƹ_WA<~*a 0W]v gkzB,RFpb,#Fxg{y[&tHo!\I:EI9 {jcS9 I8Uip}3rOAQQ{vtOq`}Q{ZȌӵ.Y*VّCx>m%HCb0˪륟8EE (4x*Bcup0&z:#-!Xz 8c5cqR!8P#5J If 3y4=;c7y>}ufh@"%9~R.Myl,t/@ )SO:ę1d`}?}sHzϲѓ mg=cKژx):xf9gLuw[\O PQ(z=y4c$)'8}I s0)`9E=b{}x`G >c;R-G<-GZXHv|}=7iX=NV>Ӎ`|Aw M=!k(=ݚlvUEPVh3YTM8eBT#R86/EU|U(Xe*ԡHڲET Ln}ḙFX2eJT"+R-"VbiH[{Xkrs#kKO{,Uak"OwX%YSoYeDTE+mgy i|볯~k}j,NVi߻ e脫2CCy[)GӭأsOnukٻƍ,U`B@ḏ@`{:meMvoIɲDIT%jHǦ*ի3;[?[31z1 ΤIC#W -Yz'Ə̧i5 b^aI< j^sS) Cs$;$GuَՁeKpгDNǶ,9O%PDC+#>x_DqA֊N{%p}hO?GB ?,<֝< _OBb#gCY[ kJLe#v%pѣ-muArZ78r%NDB* t{LdlGg4g:;Y@gZ`Μ.$=k5':/f+n:O?i7\⽨Nfj| h4xʘO1+cZd}D7)R=$G$um;}s[۪V2GyixhM$PkԻTF.G`x|kMW8COQr sv:ii5@W堲mWNVD%UX>xo8pL?if)[a{CuqL>S|lekcc*s/:jxӺ>i3qL{j;hU<%Al V)iQΣUݢR??k(kFb_n}K2΂gTgMyk!|@_g [`p-b7U3&g':5!X-rԪ}d<5ۻ蝞ЉΣ˃|hEG:xEGBAthpXp-GYӁ?ڄ4__ak 2wN9ǨGܔsV5Jk qy`boZ`2ΪvRo]jެ8gCP4d ^Gw o)dqD4s LrՍZ|ffm¹,}g8 P2|Rׇ?3RL:t#kny ywe>LpoEi&[#(^jd{ 5@JZ؉sv-(>l>|˶s0uq`325>N|.(= Oᣲ`F.` #ˑHG$y`i+$L! P`ǃ2Y C  J.(m=dꤘv#2ؠcgcR@':J0q-'GEVAs r1"±␘XyمmIZ6hڠi)5@)HA.lXkY)LW{@9,7-C&WNj}!Q]~,<\Le!֙ev[WV-F%\ O*nԖ誏K=9DTl]q[~L>f[/1:x($Vk @7)B+i8 k^S(P9Gp|8PBcLQɃ3֖KmxR U!s}+m㓌M\k >7nsq̶$dv}-{)rE9i(X.pH ﭓ ̩F(){Uhs vH&_p!3"Se1qPY qbE)8Bͯ(Kݒ)H_NV%j'E7]/q(D a۾ruϱZAS_$hyN^ظ1wnbD8|93\ht._q>^DiYO$_F?M8ϼQ%KmKTee3!g'h{.ȍŦaKg<+NRXaj*,P3ୄr;v@6Y>O) ZX۱`<@dY`APS1uh\ b|`CLsw/߭G?]ij}%jߪ=;|/7 7uR\[^:A4DCCcX<,`BF.ݮ4sE%#gR hPfAc~x8B:/-^&CwRJˎ?O%D8rmj:lnMcs~{{B=kR P)xuRF!ĐlM@lr̾xiOl;&GG?n?SP0 hzI$mqU-ɗE sq(}~Sc_gx#]OFOUPew߁3F D@o}Er!#X|{1*F0Na"XP{M\SZx[HhrRDf4RC5/KsG^Li0u1{f>8ƅ9ՠKLc5fBSE ƃ;Ӱ!$@ښ 'LEgYTghX0`x^t5$N+oP|M,L$tكʳX-=:7p1VךmΤww7%[{;z 3&2mN8QZ;M0 fh% SHRz lv ﳮH*^?[g1jH涮{t~s&ՎI[8SVQrA9YL1z="Yn9OlїxDJ)&n™/aj`,f F)Rߐˡ{M~zU].ZAIZMkq4<;۠d}T&m尗n:y2uM5vMDK[5Ypl1YkHرAVEN,UʰTc+4l DK1ÍFԍB4{Vg8ٵ)H*\ncp {~VSS ىZ{O>tM::5#<95JYtщPsR󴨽dzͯȥo;N}57MOgWct\{UɒT0agE%`oWVbWdD`(^eɺ/|ʫ_Ƕ4L?F1Q rf~ .ʐLQ Z^[+(NAܒכ>]_s:p1\Πqa$D vmCn&I#)QiP4AU*| nLÀw>4`?;sЕyL%WTҜ(6C)Z%|nAŀ-<|xaR/htLjd<)V2""&ZH0<)c"}H IanˁP*Ե:;YU@t\"vh5Z`(3*h2:ڀmT ԀDL,ުc9gES[{ v2ƶvIҸCﭮ+gi|H< $l2g=I^rÜ9Q #ln%9L~Dwk#V ]pwm3#s8{]ۏTڽ[#/ߎ 0n֊4Ҭq t_8j̀H!4E2kw9pE757Í'YN (,j$}V8kG!M@0mVDu#_yo7n]i%{Y4RD_+Ƿwrj/ʃ[xe r ޏZM2l |v ,U^"GwNOryrYA2[$sA0`nyփo,XPܳPCL\MôZ>N2:u#7llU^m:.t]z{-V|tmrQhwHJ*\ޥwۧ[6==IҕW׹YFdבm Zs {CD2( o]cyN&(Hqb>i˅= :ut2ۢfj~@yd(VNx{ 4ŋyjܣV=e[mH%[;r[@ґR}Ums2(:o;cƌLy-#chnD"cO2]?_ƻ Ri3mT7w@ozu&XUז{ճEYgl%V'vZK/2|>*OSFլK Vki$For&)np09sH[G‰uXT~eEǸ]#j]]rNi֏PAMr&y &uD9hkcb&zG6f0wgϡU}ƌuih]qf,_Mϒcg-w 9Zr;AfxY*fsr4L0IlLX6o9a;?n2A|*1WkiFOXtmZՊfio)uEv^M$V4wcp3d:]ZrJmzwfu:lO[^a@G=Rze}],K}VCɓf1"HII[O5*^RHe~BN(,)pşe50np%.| (< -PR[:/Uk(~ꤞ$ c¥&TVgẺ/`_:P%o"BqN0:xG/[$:tcUCkb]x.<[ҒxY]rn C=@m^Jv'=RY0'=:n)RviԩOVBx)/8.͐@Z߉*c2Cᓕ,A0Ks8t#0 Vݑ)ܞg#J48]yBg;l и8&{LUO+4e)0BXƌ3Buͦ>`)qQdחn$o)|sc0M).KO湛tu xXWOZZ1`9ERϳlQxoDINpԈ7Di)QM { 8S_ՆLR*0?'q O= V>mrJZ7zw7(rov=)ET1b`kn/EU}~]f~2rMߦo |< 7XslQE"?Eh2M NG#%@rͯ%qiq(6f(K` _q,aX7ZG7ij[ALN c oӱXt,mXB3R c*9R N%. 69<,FI" tc:mMF-7 L0X%%,C|.'d !y"\idvr| `_p'?Eqmj3KSa@. 4y[}j yԐyYkzTxܷU{GW嘗*mVEF@uxr鰂ΠFۓ>.[utDZ1?z,uf8e-$r5˝nԒnjQ 0ÿc6?ןG58 T<{Ql= n*O7?7F!(Z&#<'O1FL*ARNt+"ϾCB'᱈*4폄~ 4k@Ȓۺ߷M. 9?e.7%.D){JIɠG< d,e8"B c?ο|9`>rY04ēH=߉ y ƼÆ:A2I*a=R==() X ńN)PʼnJH{&R̃>?="݃)f!앑F}kErvI%5+:ksnhJ|\4e`;Ln(s^C 8 H #zH*i;)1ȵS*fs4ނ-(5 BRM$XH{M5=jTPMuIԅh2 u0-+H>*H 3"@)-%(%'"*I]a$h.C4ؤ)-Ι43:''5ҡuaGHG"9U+-DO4yćyĿsC}9Q4w{98-No7m3mx}1sL(QmGsȰ, Ƹ#aREbӶ'Kƛ_Y=V> ]*,b3=7Nw: y2GW|)PuRI\NGw꫹m~:6?mf0}ht}arhgws%ހN̵mKn֎ä XúA=u9: a JO"Pnn(ikoHdd9&Pq|2{.)˭*_{>:Oi]J0_fEF'Ga|} P+r[dX TwYnp:W3muw_!%6FCD/Ej9ܵw艹5f5Hi{iMO0w Ӑe#ߌ|s|%T8 yS,H CR(Ŝ(a| ߹>oK)nz.Sw-zw()V/1??>/ =rf g921e=rh_]0cLLqy-o*P;IԣTаyjr?ȁ2.ta𠊼)I-(Nnq2vWw#h~1@F>{B7[z}PCԁ A+ΠoXK2g$83{fx4 : bt2CWr݊OZсrf(*OǼ9N<|. 6Z8 OOJL#nl,*{K7D>(CLgǐ1/mKd¤$}g|P-1F|% *jم/2Rc3^GXVՇN̵~ր:q5E4 #JF[k1J'.:" cTs#vRr[jK=!a &LIX8F5(K)Sha=HaԬؽpzO6^8n& 6/6ʆ6dRNZ*\ܕWz/F+v$Hw][۲ѣwesqrRQ1}TTr͝2Z+44roXD]9~$7h>cnn aM TG"%1, ^( J3K` 0wwz8(\9$+0U|g0Iy~k`S.e Fv#l _ΜHgHs~_~~HҠޓ2ɏar{F /VQ]ZDC5ANжI;GEn\q䖮@=7RHWDlvQj(\-Wx B2z`b<(?Jf[@s]z`S,`3,:213IXpx)C1\yå O=cF@̇t1gz4‚S @F?f~bN; u'zIϐ) ?/cy|05I7\I*{rWV ~DbHѐLz+0uz uXe{ME( $ȴ1K1&zG#zUk:{6l BJ#68R$#7g)fb1V :a@QC6QP;Ց A_*a3+P5ݯ !4SZLS3& @GEDB Zˢ` phX6N< ڞ[AR 4-|7^>ۥ;ƛ."k2޽g}w1BcFH75jxTq8@ҒPf!roD Aǝj켳! ^}珯$^52 %ZR N9ťoW12#XՀ93p^{-aG$]\[}L-ÞϹ[/df6[6Qu FQk %I.P!`V0/2ZDJGt`ZD"rR ! QZqXBgO9(S:$0 rtmH?>r XFyV!!: H[XA`Ir.٧y9u8-GN9mcrYsbOJh!ze :yI,y{c~_.u_qʎFXjIeƄ,;`x~Svry=wFH^WXǿc]7Eppƥ4Z< !2AR1G EέRȰ, Ƹ#aREbѸ g? mH蓹zHXmGQכ/s}B 2!`; r;(u62lCU X̴Q*c e2+gl#8|ePjz:'OxdkIMhsS]vkGžT&5|fOu39^@-.Ϫ焤KpސY\zL'Z;T~U"<9w!طY$d"iY"4au  `{2#^{xIXefgvC sNme٧v"gDjP f t^ҩ9H{bC 1s:^x:[j+{'x]wZ7}nRNW9 vC|Ij&Gj]ɞtx.0ZhbZj*'0[33 ˌfKn"#Jьd3k,P?p2=`&59`GQZ?Mhg:sY@; r t_q7]8)ε{;]B*^q? Z&T2VZO.\"ͼ8Ku2m*6((dQ_4~du5))]>-t"`,yJbXO}SQB(kB&0ZL320惤 Cm1rջ7Glw!p#&c1x4dހ/Ի^,/-^>1IlS*&}ZNC>KJsHi:zUX*\ &@3o'a9Bc`-@'$bRlPŏ,j&-=&Vۋkz׫XR^5d\O ǗEɾI&B>DN%,"4`wNe橰n_hm'SIc Oڕz-zSH2eK=.gwO2ޢdaDt=4U5y8"\>} ӒH1~)0)~|ꧻ6howFÏ Za>o/0IQBǮI!z f&A[*sf`t='aЉ+>qٻGpgCΥJXݴfVa<;_M9uqeuǘ KZOh@DڦE8GH|w+!VtKm/DjղoC4)>))+fc^uO.=H-G I{_vyRC}vW )D{oHL{H_('D %d(Wz{]RNۮe!>si:*B~}~0NleV+Pp$U82L5(ǻ]3rj\fa9I4nɼ(a5. /U9n@r-W7~2NG-_jE~bJz!Uo%㑅H>HԔ1тFQ4`pHL~u|fS.Q7[qi.O/W~՝+vZMOw/!F$er>26w:n i?~Jk`-h^nLjiVy||Y3[8sd̑1_1,C;kdgf{hyS1ut_5ItߠRS0#9B%=d½QʝroӗfaWfq$%TVR]V}V4?N C>;FTl>ޡxQ]]fA8{nMjrw{b{V؛poN9!Wc)>{iӯ|_a&z&ﹿq9isd&77e Ӟ2@a,7kO?32q!3cȘ63L02F6gL+F՞͖DɤZd[9隮ǯ[{j{ 2!7ʜOpPWJFSV{\)UQE@ r=$Ty@ͷx`irQ \C3f=kE5X :i!by;KH$ X0rPn+=C)_ʦ4I) bym"@FHBGEI/ ]䤑??Ne.P.9('D.J&akhKE.VAr2NSS^8) &}nrzWs˝b~쟛 peRGt85t:8)l~fZcQا6ɬz v|qVG6\]]~WGjy[STI/0;X[/ޟ/C^6a%0t\Ow0}ą2^m;KmQJ`9~^x$/[?)هkgHKYcHEUC¿c^t?/6tR'ܿc[ՕEHw7 Dǿ1,Ziܮ8"]_VG##A4G Jts|' 2rpog8P5Q9ƀܛm:if#=ˊ~Z˦1¼,g!Co(+u|kcONnbdޗhޫ=럧EG T*:}7}px[cЩ'[;!/֎3 {KУSܑyneu1AkQet?~[8dg`x"{= znmWviA6_9ru5/4AD!5J{[r΄ }a߽)d~Z[H VYͰ'޼cxm[n]ϝڛpssv<(=opO8xvt@{.[ ߔЀ!A8n TgXwyPTv0Zy[CW@Pv $j\_byRbDR>-j>yʨ귫N{[4}7=09w9{9v>zjG**_ y'e wOܰ[uͺ6-G$ A߮aaCmB_WigprN!fG.(*IR>%Hřu6 a{ cmO }(2hn dYu*B^0J.ٟ6Axp$U lH,p i.RU h,_qNfVߌ$ggݻ-V"=-e Ҍ|}_wí {W](Fd78E2DBV*Mn!urZ^36/OӮ< <-]^lɹ릴q$[\r+l!Ah})a[t%TY\՞ ,38O|_^drar:_:ʩ4 C\(zA DJm!ȝ/<7%aɺ:C8~AV!EDMTA٤В0eH.Xsљ'BhCh!aXFe 1TԠ/ -R)A#BRO`yB-GI%mg-a  >3+ty{q.q}LifQZd!ڠ0c\.0r58]kq>Y['~\^(۶㍅rI%^Z 8HC$J C#3t+'s8{ݫ žo Sy4gpcYi m 5#x ^<Ī0 $`jRry%$/!|')~G'>{WqV^թs?L=tD<;K`3;.}oAJebQ+:H@:k %֞ս uZPΣε]+1Os{Mg봙=St5O.~!|~ẓhf^ِgh>'x8xD)ہGWl+F˵?yӓZrg[Osa}`6hxBzH=y"q{Χ˧F_tl>s`"I%MQ:'W)6s*9>Z(\HBM3B*CR>`EXw,}a IDs-Xl9l:Ù Ԁ}njF~3S=NIlDFk_(dV.e sgSc@dU bҩhFt8>Qt赏E H%8m Bp1DY@r!Gϣ~YC 2l ԦE҂.E 3[ ݕ`k).s p6΁zf1$v>;S†p1N_MS{b6)LДikw >}X4(vȣ cZs.>ytZ["F@W\7`4u.g^ 8:|(bqIAX+NFxh,ͦ3,;qe_gVFr}~v$qZo{JU!MpV'!NW_Y&;w%q}Up~ՋnvX^J&w4IDA(uъK.}g% /ދd8κyxdžS-B !u?r{n^u=b't<IގyߓVՎˮ]mh_yߝ_+{5˺u''[̤<+v.aV 7vgYשDj7|itAV-q/Vշu-ݜm>>7̻7S($ųnew܃yPB_sF:"0R_wg3RuՅjfS"c5s&pJ1!X $;rհ4!#}V}F{B`Q-Y)^*?`knus$m1ҺI,f홏Ά\ ?"'Jk9+AЄxOѫ$Ch}rdG Z/h؀W$+ :QiXY2Bn-(O R&-!6v@R֎M])#`%\[>W+#NioR̈k3#L$iZYu2B7')d!"a(dֵ7A(3D-'@Y%~MGt{+ F r2H6Ŀ)NSvF|o.Ӽv%ShMu㖑Wc [WR_~íFH @49 ˜&Htsd01YSu9٘uՏ@ }]t=!$ѽioqcP}Py?O6 *mT#Uc%$kn`'M蒒MOp2 J; D!ٻ6$W~:<"/hfm4ԦDy1}#)*IIˀ-ʨ"2h|p ;B&v M-g KMg~ZΏ&d>Eu 6I%-dyWR-J}[gTq6LEk_nk}2 RQa&3dZZ֠#2:"|N5ءsI04+շUǺX*UMqt@  2XӨɡ"$%d ~QL nx K}P e#BV !.ɚx<&,p)A_KK /5Q$G&:jI.gQ7T(r\bٔ,X.}3Z"U{ߡ4?i SNӄkٿeRwCTą1XYgؖNwu]5h5χ 6Rs;7-wJKCPx-"g8Y̠Wb#'ZD8wY74dym(b[\4yU^@Ikb;|ϥ(YVt;oETEjlSR5E3y[oSb߽'ѬHzx/9q_Iy.`m/ evFަBK)LnC|G2CMP`K-i !BQit_&5%(..].l~e/ޝ]ኃdc<6GU>R(z7Mг=-;(rGnPH"JR01l@lHM'VQ.ڄGѲ B6QodVuR_ivh<ш@VK,`Zht)Ą?T2bd|*YKm̱,|୍fB6*Z陎;o{`nq4(h|@h7RU/?dtȴ$9S d2&Mю::JP! K5s *2V1m&{pd=(`y̼2RS1"q'RQ-;qPy-j'[-Nܴ񽹄SOΓstه}(`^ZRIZ౤4LhBA=(`+̼S @pJJ!>j0omuwnLX6'Rj Al=(T`pfUVFxb񃸸7݈WfFۤFg,e5hKBŊu;+¯o!.O)h(NK׮ن(y|} FFl\.Ϛ l}ϲߒG>60|vEO7HRR2sC7̦=0cp[@De+z7-nSs"FTD ]N볮p*^M@4Kh=S(ن$`.+x`!G\Ha 0eZSjq6v⋘mzD5/6M^Dᛝ7 6J|u<[ nb;j1RH:xS9V &cGNQq1UtlvԮ oObXD,,~čF/|X5@gX࣋ e]J^AF-Rc3RAkA8.4,K]F<|)A$ 'n½6LG|2Ҁ6ijhˠc!Q4 @MYi&"胁bZ$Bn - dPCZ 6h!55JὛ֢AD*w.Wv\Wƙ-MhY!lU@We.? ڀ DypLXgLRZGЖ`d)LJo$y? )7]YAm=qoҘ}4LM((j s{  JK2,0gZ` H'e Tb2ʼn*о(F Pi#>]SS:LC*:eY,pBтK*Wq)B$ |::+{Qr/UJAσ?c-Z0>|3<A_Y{ȉ/%gIҒ9JTTd=D(ہSl)&1)og*]F\NojbR#[7gC),KCzJ+=ZsUR@ME ғPq#*.(,E.9ɔ*NjTq! ºiҗhN Yi`YcL{V+GUWNwT5 FªKY8_W>V.}WGX; zFAowK{zѥ-z{;>[Xj2#1B RB]u&5Hѡgl`L2և WbsUtȁ{һڐ`䲈"nUxWxrF$f"UjVFզӳԻ@&׻c†p>]LןxvlwN*}o%"qW1_{5&Q/L)J))Ř5^fV1ZtRLpYUҔe NK\ 8}b>MNi͹ J8Tզӭ;q<x߯~ ݾ:"mEavb0v~k_Pz2/>#n4nUc/j XE>?<'4a;{o#j[N7'{u0/@c[N-탂l+~%Դ-o qXdD`njjd4-3«dyI%߃fA1|=ܲ mѫWVA)^X);x 7{vV5Gma ,XN/=$l>*V]z,'(a{aIeIZʋ˟E rr<_3vBwO:hs8NW7uLh9SZ[]5nly@Q)rM ~ʻX"Vf 11ne{=7Zm6$-;uf]&Q k*"b'hp 'SJ @/t}{=孕m9_G˝Py#>z?ꔻ-YFݻ֯ҴPXvR? 5榥q@w[k]AÇK6NӰfmrOz<74ɝo5ɽS[ّyq7W2,FWms>^gIv}@;+#C)~Nou2?BO=;/kwl;?%6K4W쯆v׉mu7gUj׼DДٰy鬸1TT7W_T%84j`9" 9p Ù='t[^AL~%Ѥ!3cѦ3j2uW@X鴳Ag{Wi_aCG20]$ٷ/v3xڊeIg}òvK>dWUeUMU U>iԚ8QƅQg2w]e9⨡e @kcXG,y:I^ n8[yar\5n.vm M&qgrlV5XE!T,jnIKH_%Ί@gBJ\k)rm6 dcAk4c9"9d,Am"P~l < ʈzn.CB$A8%0YS`=&,p)6儎AX[>@7:jI.gM0 2xP`= J(f CH2Ҧd[ߡ0-\OVK.cZ6r~υ NFm5bUŬC[]MA] 黪NLآ{W2_}jrsf4e6]~>WgcYCG S VGVNFS|*ҘX+ݜ'ܟ4o5 _^͕Y‡Up uNAeAb(?ѓskuzI[27iMDQg.PK*/uڤ5\p .d/1#?HICL}X."rG({|?kbζ_JhF341)O.퇙ud=s)sQlСgKIWÜ+7B-ip5B rУS&5ISt4}wQe~q?-;~Jg.qScxb U?ِw=eг%w"7qV2v$tZiܯr'!R .9zi6!D:GsCnVȆ4%As"]f4ShD %fY0ۇZr-h4:b*xRQb2쥴ԝ)M!ߍ~ڢ;[vvG=½\V&]hZɃ~G&7{b}XZr/y 2'01dZ⥷$s}u2]T(ѳ^oVw@KkBJβ$S9LBрLIa=O,ڳ^y6H޼'!Yr1 Y_Oh0}r~ ^pSRQ~gl z$&Q`H 's>t y/6\QGin3%BOP+b]̤zS.!->!Xz 8,10@GcM9O8QQ,@FE+=1qpX;8lٽ0nQ_(ZhњͶB끐fԊgpq ~F95㯣?[AJY0Vbc QmIBEv Kl|ܔZEM@j++DU|( LyroRf 2۫Yk|`FcDlȞfß').T9*<BuV CNYLqgd.`o7OD<42LmH2KV~B|`赦&"pv6󴗝y~^W=6%Y{ى9=-ʘ$‘R0H<Mt)TB[:z@AE"uNo9ä_o^i~t2ˆ 8]Tx/rA(T QsuFVPEP) "ύa W:R₂2Q&XNdF:urBJƅ.RYJ6S<Ƙ/u:Ej[2MbcC?/~Wzbz|tp-ԔRad3O7rzH,#6)'@ !T#^#K:>)f C< 8=yf#D9$?!S`)" \dWtAP6 uI:T3g7ZHo*tϙW9v9ކCi2 U{;bdqsz85"gTJ QKrd8A#Z pD7y-^y^D &yeG<\ῃ!;kv*4ѷtR PC! ^pG\o! q,k8FZÝK2%" 4>[\>Y'-]NFzΙkojaP!F)bI![j\j+qgR g̅ϫZnk S=:\ӆăOdJci`p^|g^cYj~U{#f(vseԤK FJ#Q(c4EpY0T;mygXXgiz cY\;|۩Gɱ\@>#Cѥ|\|nCֻdi LHЂ:[}1N| ;_Kh32p9,:m12HzR/3Uxǁ{R#xʐEɨVq<9#3@Ce[w g;cA[PMh4i7}R=?_BwϾCl M,cbϓ#*h&Q/LEKJL)Ƭ6ȴȨsRL!p*d;eLYKIVg>1M)cZs.*Zֆ7ITnqW ->lN1 ۬Wtz8WGӂb#e0Gӆu4.P INr;N NNg%Jޖ9pb6L¢?bI dDCx<%V}nmxx&h,uz u{nuh0\R'q8\)^C.lS<&غiMw9wth]l:Ė[mݝ)O=z^jn69C{BÞwty}o8ܶ9jS%JuzYQ~hR(&yт:t"AQOЌ߽ ԃJD7&0R JY%Wq:['M0k@Z#]# n8|o[ n/('<4XRzN~^^@Ƞ-KBDrpjh|h<v*>ʴBE^x4lvb,˷9>۫W׼WyN>0Evۭ^719 M!f^9:DYet|"`ə{{ 9| VqX\Ѭ;WXS)n COܣwjc)[A=hKm6νdKm2.5r/Iʭ;I8:m\G(c{O-xs_4kx!͜hP=BKa`i,l}wgNPii.htk94|tƳOŒ΀PTݝ8y[2`EAo##T 3x#~D̹Q$aEP>hUPeJm%SjQOc7uG;t]|lsǮFMֳ^z1+Bzj<'z7z*3+/gIU2l*Yn"fnҤgt0êu3.PC8gggt|*K^R}RCfΪ58|8TTIQBWۯWiӈ/8># ㄪ3ҠsGϛY'~3Qz˰ =7(AJJ{b%R?`b?ޛaԉ#<32?&(Hfp6Ueo4¿f٤ϖP77oW}OWӍlv͜{~. ~fr=Svj_8ԻJRQ NjZ (l7ao? ͔!<gAITdOyӽ6#8reN`c$1KoI#p^'x$_ߦ7GH@B oU VhA@QTx{W-'sTrx* J#bV CNYLqgd.`ga{%XٶY.k$bҧ4u#nwg-w~%lNKn޵q$ȧ#% eIv?\|A?m&ImpV(Qġ^ndΰW5տrڸnu4+I9TT&!E Z8ݡ j4*3D&jf@F&0L'juè~9Ƥ3hPE+01̑,pa''h4<T)dɊk*yB+H\egrY|'H>Y87Qc0h o"d}de/F2BVsi i˜6#t@F#0J  URj/L5%Z_2h2WTq6y!݃.%UJT*]RKtI.^^VU>T*Sc|LT^T1]K1UìY*S5T*Sc^1U>T*Sc|LT*Sc|L1U>TK헥xKFl"1n,6h,q>2-g(kqZ\ϟ]qC$vawڅ( SJlv%ӘjN8%J(,-Sj=VYhxpR!8rA(29̉ީPHStصYT;"չur]t1p6"'M/wq|j[׎1LF;[jd{hw+5\ c|4\d}]bj.oy݃_s{ﻞ>bM ބg/C#?MltVv/"E予 XdB3Ӎ/BXz":}Ƕ:˛0x}s`ylO.A2-Rvd[GAs3g,)oFcL(N=% L)B9CzHhqbX2^)U֕cCK+I~:myyx0KˡC/_oy,F\is.*ͤơEm ]VTNfUAdV+G.ABԴPNKwY(Q2G0V&CTtpn"Г!X}U(Rܖ 匿5}_=Qz#lC#9`!K4)Yƃ-9SP} 3M9a笄VC \`QK,[䙃Yԭ!=ɛC/;/q+AnJZ++.Dn$pBhZI(g?Sj{|U N4XZ^s{<ʼ2-pZ2 K*}TT:&հlTF֥/]>/mf#_j:%WsSlۻǧ~Xsƍl[>sWq>]捩ƥvruc0Wƣ6׶Qm͒l;,OkGځX.4On #R\%gѧf ޷yu >QN*g:ߜg_V)€;(i3M)YG~tk\Tg,{sv߶eg_Js~ս]Fh{0$*R2"pK^I>x8wJ.㉟N< .O3ZOpűOz%4vLdӵBt% [oKథ\]=G>;5<-Zسx6M}BkƔ:`n.Z=ܞ}IrSnzѥ܏Ia ?\h(4ӗb6yz2tu)~gQeMJi?-;Oqh{l{+m%];Au^~TW C\D7z%]$lFHMJ/T>9$"txEn-MӚϗ-=3߈ D@VC''( Y# -q>妐JZSݫwi:掠G!Y2y -f_NNn l@E^vU0LZ[hב-mEZO0Eħ*V9؞6je{(Kk8y3e)cvwWUEvA0.޴\5}&pIWŔ\oϧ`@ӆ:=k[<,et@h\.cU"Mz"<r6F ^9&5JMQ! vp o*Ut"tK|@FN(巄^DٍfR0qHgmnR]r'@σ#=y(|&s:uitvTKŹ&C`A[[v mʤH@+Gb-Ky{tuY/.Jr |Si ۅ??Ӡ?Uiea> &CR ,L&lϬ9bf.|F^@]e_eVbʁ@xhȘ [(\MK̇Ub09ﻰ jc^9]ֆd@bH!h5jn$s˙&K9o6%\-ZZcZڋ#E/"A_z^-8Z[O7?h dr+ȍ6L5o(W`Uֵ xȅvru߶䝣,G42III2BD)% 1Gn􉧳Jm:ԻwArCu-p6ig?&q2zNkc338lA9o4R珝&?n_N݀ltR37Q UKRR)+}jn< SvP1UgX9k(Zs 7B螽 )dB:S. ti@; :XTP<V<\4FTHr$Õ|Pb;-$2JP0%ӂyLQ*ŅG<& EuT,6fhd#v6%ft]nn|dcMW%\AMLGfu ?;ΞQc3>_ >i>Oy]&1 % ~|=KL+,H&6 1l0gd)dT`!ڢS PV`8sE'EN2NwYZӚ"F5P5i2Sg{wA_=!L'cQ+oo.VG_^{}+=YXI#Ҵ٤޸Mϊ*kf-Rџ/?;OžM%Ut1"ɂV7'Y~ѧiJTمK}7ލ2[s~yw^؂hA ]7zL^^^zL:_45Amn\< (CKtz[x4/fBD \cU Eܺ-mFxY9Z5wĬܪUT1 lN2>"Ĩr~~l^y[;}?kmr^7_p/2hI;8xμlbmZSY\%cQ?8rx> +&-uxf8qaT NhiQX  ۩U%jvIjR"\xH$" d%nbz x|XS>D!1f%&f'[Xgqqr)E!ijB"ﰒE!W=+9sjuۚtۡ9Nxq3J]w_ӠY{W;R~ݳePg @52aۢahQ&  m"U*TN8S%&rT}HJX7)n)6JP+54rgx`TOx }zm7((APƑKU6 3?6_JŲ:yLf|sV^qS q_:?0дE4 ]{%+x{kr>b1J&ijb (w[ %QM΢G-cDMe_E!Ku`! 0g $j1kŒpEELY-^x:VS+5,{RL"i  2O}|DZi҉"ڗ\j8,ou|Nj958g9l;2|v>ml`_=dvG_㷮%vg=囌lI-5=>?Mi ?;V)Ww${:Hs8CZcƠSbT"G؄×8QsׁܵQc6Wjַݎ⳯3{ǟsn^4E4O#`Fs,T^c2(fJ6`d1t r?e%!N+0 󨐙B$^ȴxIE IGem {by7cFSBQ&IEyPRf$eTOu"7nyKX紞zN;&j;S䞐ZDx ԗ伾hrw߹grR3v4U*].m|%|*1Lw\ KzWZ=jꬉóF6x6n'gf0vWS}!cELٰ-)d[m=gM˗͸贁slUu:Vo<*Trv\dΪckqɮc()ܵ[@"b."v)j"+]!a c)y<甭u=]x%'J)=ǧ 'CщsZRlj6&eӥى#WֆBoE.Y@#  +:P>XPn}/'Ƹ*y [[Y%a[ oW2~% -+isHMDbA7;_ʔX-3_BZYl|-[( Db EQJ#E=C+-K9}oCƚQ&ΐ]u[^tJ&(}^׺OC{QM w^srzKg]2RJu4H);?>M}pTL8J+W3[IJߢR/RTԎd(bt>M@"@r!1iø%TT¦c rZZhQn qC 4dK hf(%3@p&e66 s,PYabvF8Ev|P- c%'N=c￉Q{2R'bER69Aڄ%$QRk<$#1hM63o]7M}ՒgIG͓^*8cnh:;mߔoāMSӞ\ u!T0}CBBU u ^ONw|ڥBp5 9#>4"rLIDBÇL0T#GLI$#U#r!(\Xύ$!\4 L$xZĜS)O1btcDՆ~#h1ڝU ϞS#?}(\i/tw$j>=1 & LJ4Psl}l|/%8 !:U@R4WT2v=g?/~a\bL;ZHC !`O%q!NF:(KE'x8) bTmAߊVHu mm"0,6+Ld:-\3 ēz2iדN̲Ak6SPُ͚ :Dž3!1!1>VR(Bsu&pQQik4ZcFxI.Rt*Ř[d-ņ;qP"oj Z}6wMv,ϤY8F'i7'Bq]ⴂH&"O!-qU$qkoû}^{{]o+wpoIW3T|&{RΐN@2Q-3%}@L@OJ 8Ud8@ H* DU1DC= Nt֯hqToF>7oKhx!}GUYEnƆyBGg&ƼWʬ!=~qxawNYvXA@'=z.ZOԸ_?x[ ( E1o!IǏ}fD3QHNT'C=x7)1fH1 )(J@E.f 88yadpA :8=L(a˪SL(=BC||؇s= *Mm%*MyMIs\yxeG%ɽ4γ\L3kCDsEt2܄hRIqZP,5(vU1H ݛ]9C! F%S69k5I,] @F6׺$n0iN3molk-ƀtamԵuaC31>d^0U\N3 ͣ&6nk{=a_2+4vUӺUa|bd\n-?=:-v_,R7o`JsmY"Ms=V)/;C:E>F:4^7?nN Ek>[n-/k5b =,1\? T}rD Uj]Qm$"*N{T٩ <!9=I>* 8LCbGMO,/DYSk- "urDzpAb"&ygL8F|syEPkf:1=:IkmF_Evqi@pwHr{`M]Hmg4#ɓxݯzlnٲ$]+gwfR]{(7~|ޠT hٍ7@6V / &cL)T&&d&H( $KT|P:夜#g7Q)iEe_.I{pmwBp^KO&e.vnh]uH{î6"gS} έW#n\toZ3f^tno{Sܩ.x8%D R/ n Wl}>sbQ2Gv#D BLlV@"D1hZ *JډbOǾfbVٯ=,\i2UR{<΅fw4feZuދ(sdЃ3oϧl"e~4?41'[UGZznļzG(zG(zG(zG(zw37txيlcu 4 Y$nw3{Б3  G@.S)&*v siM)UHPfԍ`sZ Rs| B"ypov OK۰ HAKLN͙نT!yh+wҳ@,\Po!Y(T6dƒQhK@[~3FQEՙO%w P,mC K`V$a\MUа/db^pt"8gg4֟,:mޣL-୊e*20i 9 Pd0rRLJղ#"<15d_3y(鎙Oϙo  tђr=WTB芐f+H0BV<[wtF!;-A/'U&nN!M;dv)'ץSNZC&,dgJ6QVh^gӚ5ipB#sAJv!(uFAl"FՈ¿MbC^>:\.VwwnG}w^8ܺfV귤WONCҧ쑧p g2sVlؖ6 ~<Lp˧pTETaRc#6:l°b]\yB۱˓ZH^w-MFj0yvqY.X+V6eV8U_q0}G#^{0.03&38 r^M^[ xK`LW˩VTf`.|Z`b6*vZ6zóx;'[=~E[voR6scxxaVdY7=% >YY |f?NK[B|x$gT7x=_>_wϙ7pfϘ=c>cy=ngf))>z+ $BTUzפR6\ m?Ɏm2|u3^퓔[ aP#v8RVss;: h>V_q߁y}tp;^nS'4)]%6h+-`1y;_&yLfd2˭qW[B..qoTZQu]X'Pk,vųMש>¬IawQःN eL:+m;AS(E !2KplL֖nB;g^)0 3xP~\kͦU{7)Z/S:Z_|^oGJ% BM KuX@)Po#N]m r-js_~ d\ }HEz*^E : laJgk[ߴl:Y`Y]'=!IPa2cvA̳zΖK:'IfO7Vax󤆳Ffrv2J_l3eWze[!06W| 5OpvH6A,W{j'.fQLIX焠_.swra@Hm^.<׋6|p[y‹/5ȝZh|9`ccMMJO3>ITy ނyUMO~ZVl?2zO?cp{̣m T>hNg;܌}[BOAg+eQF m3ԴF hR+iAS@ f-$¾w)aLHd6d-?"LX}`tF9-w)R6ECTNo6Ҽ_]JWҥ!Tq'Uzj7zs 9"uKDll[­!XvPlo`@Q.,i!2D^!( h/ThvP !ɤ,W lר)-3E L6 U 28yn&GOt6YNm+ΪrUu@B:= MWmřYnd8{lwWjL& GӸK\b5z_+KJ>yr QeKtȱcGBÿUs=c܀[4;ؾ;'{r5Id481V5;=:alK׌uk{kWb^9 eJ7؉`RJ&MoF5@#JV44Yh@8 ONW=H*t)8]honv)3kE1zg̶RrZs4H࡜?i}[0+d˴mlќ CSR/RMMM{T!&Ƚ!;:!ca25hhU+ j6%8RKkWZ2}̷7|iV\Ru+&vwE#,T6>aALziMTBkkRTaJ%hW*ʻ@o$G.,~}.`}Iqa3i!Ȭ1P@)bTňR).T>J1T(/2]3d `1 F&smXm91"Y恅 hT (&hG21)(N)b9| fm]}6/,͍ϓɸ9Q#]XlSc ldQ𳇟=@9Lfy[?ձw_}kg1|,(b}jȒo*.85bF-EUM&Щ=BnjNj4!v~rPD"ό?OdEHcA>欂 hTv|)~2I j/.7I%hk i>Ƥ})A[5"L00 xm$kCj9l)uV56*f,tHA$٘y%u] ᗜm8Ƅ2Un))Nq@QiȤ16p O2φgcV%1+ATSkn4׳U1IcAX>N8/m 8ht\Ș߳RhH.0A8aѐ:FfE4VᝬgՖsuzLwp&.yXiYrh@0p2jY.*PS_[?Yv.0ÅNyX0yjz("V5g>N㤕v!3aC3(dz#I "SfJg(X.1H$b8ZޜJGj2BdD͎B^##):1|JnsF2T*CuiCǍҁ >C Q0%ﵭbFv;ŦݞU*|r|Tx+@&4 (R1YF|ZЛ8]eӸSG_>aXQ=V`줚UjTI۶8GiH i_O$etkIJ$Jɲ /6ybRy ~ǠR˥XJvd)n ;gSz9;/i 5wAݒUjgZj?{6@Os㤬&Zl/URE4ql "k,0({wݩwʓ~ .,TM'#2_G_\{S>,_Z. mg'9._qX4=s>s+] Nϲ#/)yYE-FoXY3I)bRĶ=ݎ-OPiYE@'=z.Z5fW# ?@'|iJx|>}kWg%_=0h{|軞6]ƾk4ߚ[Zo+~~](+5܁k@2F`Mp>l'˛uqzs)$`]:ѱrf3嘣@:whURQ}=tRYR$q xmRN(+]B2Dls`a%D/?m`FKi.*ͤơEm ]VT9+Yf|8agז> gm:93fqljd ]>2#s:(r|tG7r+Y'\]=N|"%1DK|@FNI)巄^DfR0XVr(^Lu.wI tD=v~}bv%)6]bk-VuHћiذfL) *+_Ԫ\e $[ '߀ccK!ـX'urJRL+F'jr mqH6 H{;(~.vWTNL$Jjۇ ]^M'W{7>_6ĕ1A]t JvX\ 9e;ĕ)D Ek~+BDњ`B,J&OG AtP!R,B/9T *d-Ԍʁ)>_Р,5X2zO’EO( ۈ^{ۓ(\5v _4/?6bnO;?~_1gAj"'cu&甊R0hI Dc\N2FȾ i` Ҕ#65`-I1e,/sJT6FpGe LIڗTa훪 Nʧ̗l)4 ¤bsʧ;>srE,WW6`<:ƝDHh4u[ȯ5S5fJi:AS(zEch}1=["(Tl._jM ZHEhdrUx[8ZN'|1eHb*Uԣb!ceN'vAQdB(_,N瓲(֬`b4oۺK!xyu)噁Jy QSŭk%#!6AKNieC3Pq9 Z>2-*ƃ켵11Κ @^^<;JkEAHAK$'MD!UHL5Y]+ŊP"uWvEM؜( hL(<]m x77=GT@Q-;i6flRz*`,5 caalə咀zX8,*LT):>"KIif*ǐ(,@@倶[-'_:S_3y,tc-Kg ^N:֢;-WZgtEHfs9!k4ÆXLBhxNy9I79^AHe@lL&V c|fV㵦y^m㉹lA`;.bT(7Z ΰu᱉עU1=)W>>?2U%k^">-N&_\K:py:{9ssòW ?C&~廊iHZJ A*5m ?`m5bz`5}ȳa;Un>t~iwl8#.Ii+9v>NA%gXG;FP'k&=}ms+] u@GҢx2x-Z=iWqtx=֨\LEUcDKu8aŃe-UZSu-˾3ԡj[ʫbQq7pyZFұ6%K@:G (D29%8S65'PM֏O|zt^bg r]@Yo5!JݭO+VȤT+TI_)#:W x+F˵ 7-eҳ5EsMYc"1T"SpE$Ac0܄kT(f !5kjgg% 5>v%L rDDh7䃟0JxJ,ݵu7oQa}ܮwkzG=-G?Y2=)Yahx~xswYMM`>YO{{4]PY`?mn@妖Xd+JP)& tޔAr"lg Nn(Al|GcXzM˓ MJ%hYTOP*/(THY@ZE'񘰑8'kGDx@AD>jmMYP,G s΁.RyfK\_o-wzϬ|2:B{r!ЗrHkg!!x.Дs^ǸϭǔO&AGӚicA]fuZ|(&RbuFry 01"22զ"$P<ł{'uDW{y]TQA+H#x 8rk"4C٥UPB: }B*t>=&b^} ,§L1KUW厰y#M7UZh2zwGⰞG o{?cUY%v7j8:tֱRé1S&k mp* c M%9j21E,Y7$(EbR49W#n]:zO,a+x%e o~%6.Sc9 ATG+ċz[{cQw3|q5鼉M.{ ” ޠ첄ECWQ@E/)?YzpVzL@MOhzI/!=f #QB&0ȬcHr}T 5:%KzvmTː3V y)~yU>QRF_/d4vUE~eP}%|l4kwSo{ɇNS~ga=̩%0X]_) *ڞ{Nٿb6{V}S[Yww7$58=5S5@6?KWx1)J@z%/y}KAՑE(ZǎUX)7лWE [/5@#1 Be1;Pb 䏀)(] kC+`{mZѧXך]g^(ہi=Ӳ ?rLs-~+lO;s X|R8 ̲TB]5$G}&x(< 薗LWTeKoϧ5q燑 y u:k9, oXsry\V-ӣ!^?._%4QH٬H-[0P! ITM6Hme`^jp (57KI7NiiI9qHh׫ͤ@<6>s:iP/jUy']5]mBzD\ bIEOY '$N;l ˶+Tb5z_)^z2yƥJ/dekQ‹T땱p YJӜpfX?^f`jޠ-Y̨5>x} :WʵXx՜QiUWmJS!֋(- [1m k-VuHћiذfL) *+_Ԫ\e $[ '߀ccK!ـX'urJRL+F'jr mqH6 H{;(~.vWTNL$/܄}Ag|^ꌹM(]@;a=vl& SCXSw׶>*>1'zR='_Y,LDh,EZQ]o[9W|<`w=@1tc>l &Hs,ӓ":bQHQ|EU@L5іܗ oHTHqlU,A+*>'Mg>$<W墠6:.nvXQIt$DGpaM]!!$HL׃ %puMb5IhBf +*HzNRޟFHs"iwA҄g(jTJq-,6/\({v DǙm$%LB]?fobFh{ I}N6E-cTGڛĥTL$А+f^8 gRXfA`㨀ln3|Ty$) 㰀KD/L.]I˿ q.z?,k[*=_7zx45 ([whCE'˂KB_~$c-}3fN:M0ѧ{;׆ 7p5־ hKi`h`[~0{0j՛ 7^]@?͓x8mlՕC+D8?7#G7{޸^v?٪PN,\bMn=_mrT?}3Tʶeݏ84XsL6i_T7vMQ|չ!LSf1ݩJq,htx@ܚ7_]l7_%^h?q^N-~'R j*汚\5?iOZ5?iOZ5?iR({a1x^( 5j͵Zs\5;]D5j*c5d͵Z%JZ͵Zs\5 nE{ȵ:g,(q|~+Ø@4Zؠld*L^09i'ҷG_w+XC&ڋ3Վ[{o~bOK '<ѨsD!!8(ɉI>y&-7*kWyN{Ί^ < 2F<j"0 O{) PR9[z,dBr3%}9a+ǛqWd!52FW,37x.*B2QGfuJ!4W1kO"XpO4w]T˗x?hf=1m 0qȔusW=\Qkkc9yk$ym(Y?( E?3S-ٲmP7y_Rgutֿύ3>ycgaB$}!Q;G:oMWZp|)\++ :{Y<9T8U<%n9G^F|sH9a%|BI <XC#w&@)Q P)AMPXm%ǽ n>|g?1w}{mS+ϛ;Or\7m˵8+;gG{7n޽(xwzj z+خ? C].i i1CEz+تEzkZEzkZEzk1" I-["Ho-["bRڟ }xr~FhƛnJ8, 1VEm'A91BE;١l6tv0z.'G?!1 iHLsF@L5і-/O#bߐ^ Vì"8u~t+U:mv˦cEGGBG:?+gf-T+C3_J$]5E'1J@f +*Hz^|Vos\߁愅g樹(8J)8jwyj4l> /?Lho ѩB 6(+D;0.1j|-ʃSɱ"13ĉ+%RQcxrZ+@$SBe80o5Y`jDzM6#g$Qcg}s=}Z>;o;"fcܮ 5fכYx;!^:Dž3!1{JBRZIiZ!T 9:3(*PPFFC4R9XV[*yIgGhR /EJ\(z6h4M!.=[ ʕ=Sꗭ+fy$͒f96KL}az d#,6#i4TVS#G#鑻8}-˿Z?hwxbw2.$3Я/y$ʝޔz=Bs yy*~rϝHVdnG>ؗWnO() ^VI")* "9;RU Vd̢-w76 g]r*Hf*{uru~dBq`X[;L+XzӫıMD}ӑZ6fao i[b25z>h&#>N.wJ=xA Gb fJCq6 d4ᯙ rI莏+5uIs7t}l86ֺe{&UVLPj3{c,4J7`,k415%Wչ iVIr/Q&46$M4W@'M*hIxe9>Ů*F! ؽ18o3HR Qd0*QYíN*daZ-jg{(<4n<:8ϛ38 c[t;3S6Ά JcvC)}ףiۧWNF3="ג656֐zrd_2[+17vͬu+g}.wnzqyK-C^櫑0mf;u.L^)w.:+Xs>y))֥g殧i.} ?Dlwh4I7EPC !%RE6NIF+ߟΊ:+'NNgYZvxjIjQiNlt}J1T TqO%QZ8*D! R+}6LN43`2'StblVz2b7ٱ9/WVwʄ}gvz祯A9|L]^&GAyxGP\tM3.y!i$W95N`!ڤhR:0E0'dᒫ[Lrׄ?s rv0  "$ J%F VLFb˖$YglU%`p{Qi)]rV-嬕/$fib6Zè9c#&pj"i#C׈! ^r.vY:L"PRn_ wwvnE X" Sq+:́PAՓv)!6>04$At <*"B=u@ 9I5Gkēkq*Xo'osL5E\b 8#P%QJ=aQ&3b%U ZDqlr$.Tx;'3%w>/_1zXv?̮z|OGov;Uzvoch푲So߾jvX=Zh ϵ=S#mZ`dWwUm//O*Wٹ,JZ,~{{CgDlQ͖:Jw~~k??Am`ЏP{U _jv)&k+7:TGCnyꥁ4l%|wrbsl}-6>sqy8_a`ͫog='j)N|=R=5l0q6[ףoۗ/E0rS7S͜*9#\!'v:4ΦZ/Av$Y,6ɆijUܴ$_?/6!3!_5/J7@! xZy<-9KogdpPj wV}aۗ6a`{8z{t_hmC5 zV//Oo#I9;r f?|N ՋgC;-S4~3֚OL6aVڝ#@T)ATt.=a|jW=0\͟* J2-ŠC|Ϯ7[|N?x'n0( E`~z6œH< )pVǚ0=?`m~ͪ *j}n?([&PQZbvJIĆuVJ*VE7d-9i_skn~Az oNx -n*8y3Hʻ73M,O n3[?8Ɂ=ڎDj^EB݅L@MVp#ƈ®$%߾<҈Zg{Fp;ZoRWwhܙU;]o|uiWϐ"vW.˩dȷ;|s&կc߯հ;ϭ5CivUkf}KrjϘE~:5_ m16ᖑ<2pn?739p37ޣoVA.bdf_>$#C d8]EXKU ":7D;*Kx+߿y?ޯY{d֮ Z:z;^\g$RQ7%ɫ)>XG%֢ؒh13;u%hi :TJ?.ԒURja^R*&kŃES8#cch$kٟވR ZA)µhc9/߼֘sI:Xi<Ӹd6I1$|I[됓-pRKL(ƾ;B#1gi$0:Ҕ)bID9_=|W*&6&NE{L/%cM :Sj* BߥL.+:YVJkmvߨP[>:D!?^W*=A\/NO$MVdhiN{uLI؅(fW-~`'gRwQtϪ3E5pNU,U1@E/=`!#57xWQkĎh_j%reߠ |m 2 :e%cFHK6>*2P %AU,yKBL d-WWϮHŃVg5lsTtɨ\U^&V\ L>y(RNʨ ԳZfPB BKՐ͗RnSI%2]0k|`")4|BmQ x5:+Bp^jvqPTOLR- Xe\ė bdunXjFӘHāNqlLVœiL%@Uಭ!`oMʣ {ic֡)CmPMFPEkA&&1c{1)mp ΍U"\ "UK*kRal`!jE$4vɢjL?aZFok,`B]ѲF 6AwSA`W0c2Zo9! NMgY4& \0&S%sJ v8/P jXƯr`VdcD8.Vj!TBL;+F@(ƍ UP4gIywD<@7/:V* qJ ԛ~ԕK(ޫ0 ]lmߴo&SX$5S;t{*8@>~9XvńTE>b(8Yp)VrvH'! Y!vf֙xaO.fƶ-eͺ|϶Aw ӨA۴Z$t>^d:@X8pi'`Ӗ8QI 6RtY\iA,-*#h`1yL Ρ&qcC|ѤBB'ąsyPKPiDiPyMFA>xK!+]7C~8}M&!+xF2&zvmփ^Ls  QAwAK|LH%۸A׀R"2[Ex+4m^@͢Ta:L /B!s%KȓrFFV A}K톤D5PEb*C\ 7"2w[3D&%\HwQz u %t 4 DC`́z[ei^e*Oϫ[QW TK\ (\uB߁Q' ># U C g((CcDi)Ub.Fbǫw:P)Z `SL.~19jwfdnR4 $CcMZ bժ&6ϙT d *N՜ 6uVO;Zi&DӇaL>ƅA%V4N^'[`Ciqc&X)ʙ9 X&[ -X΀zBX) a4SQAj|sQh&՞^W} ((q˾ % $_9ۏ\RJCqc=nҘ")[`Uj4 MD!CN1)`lRx$CC$(q% &Jk|U t"Sޒ`j?Ưz.j-nZ./v$Umyݩz3/sԩܜT*%Xac9$g,קk{E?IW{[Zs㡅tO{w|zvhU,E?a/`t*†ܽ+8pkg0;y{ms"uҗ^ٛ@hQ^E7]oGWd~wH\pncCRjE%eQV9Xki6{zz>ϰ'=mb~Ae?𶻛ت7}.3Yy_5?Ɵ_S6 up\*rZsNX8[)1Yʫ~Vm8YWܬ3] /yXTJɖ%':RVpM[!Kz/XO;\!//mKZCS+9,762{/rjU t@yIK}~?_ Ćm,J;^, L`Ud^M9 !KL$V,VQeVM_&PMԛ׬GY]=)$cH<)2,edYKTY" \X勊! mOD`l7ܠ<ɩ8_59des;ַB$B4d*F`Ioqg+Q!QӁaEXcMݞwTcCUAo&F?8掠I5iMɒ._n% d]$ :YKD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;KD;vig! 5€-=WC;ZOvF)5ξAY qigP?`3`F !mګL#`Yh MA4 B><csv̈́G =u',ҎnF X=;/zlӌ?5dc NzS#yY -A3k,V0j&/a* As(6UdȏFBo7L2G0ed^i&0Wf𿇃)94 G lOrtE~>Cz8Z>݌+̪O{oy}`ϛ7AA Zq I hA@&4 M hA@&4 M hA@&4 M hA@&4 M hA@&4 M hA@&A+ }E he }- hlO~ hW˪? ہvǮݞÇ g%u4ũ&8y*f--kfB _!/1Ûy`PKo5"~bܟ^nJ4޿Н^Y>%oxzsC*!Z8ݗ.vlMMqX0F#>S?}yҷ\] w_Ϡ6 fE2ߣsF |o}olZ0ZiYYooNojߞ7;||sopm&ū vεn6/qSO[g > ,.&/k5y:a"Nqw0D>oL[Lݫƶ[A=jK6BCG n[cC7z\h&C*wJ8tv^(t$4+MJ󟛇4o tܬcͷ{fɞ>y;]h4w{3)My,I9G՘̤t]8;hϿѶS7TZ٬q53CK|3kD(T57缕N*/wѴsαqJo rf\8f1O򡝜0` Nm҅Ymцm`{v]5x =m-t6~t< g&j\k^xqݺra+|pf1,#Tdl@eᭃ9Yq xz*(@xV6LѺfRRvjckedE'1@ZTd a߱K=vzI eٛݪ oY1icWDdg\FȘ*rZU$xZ+ 4l9r;d_kk:đ=_h=?MY1 {qV8]o/xuw|n]zVN:t5IZ&8s$T0J}w=?L+~NT%[{srY7g? oL1.3mtݝ-rSeѬ=IJpeAe?w;q2?u >~F]o0ޯ?#Dd<?_~jr Ɇ\Rz8₿ba>iZaM!zn"x,A<#&Fkg~2l2MYf~m5,{%15vX>#_l7/M0ܔ~U"Qp~0~ w֏fx|uC/xo0zszeĥI3S]T|O E^71b+-G% ~w8:'!R348d~[I)[ GFfK+ k*&/SnKRi]"o)S0܇F4 2&ac5mx>!mZ8liC!k?NΗM6'/\?qF[fq1&cJD߯^SʵU^;6% dJxP5y\Guj(U,ת6./:/)pUJ_O/:%Zzb'h{r'*/n,deʆy"r)rXs,Fأs3-[i(Xue![qe6+(EjqZh*g] ]No@|f$Nt0> ?7bXLe+bj`2!hh0IVd dbuE(3Esmkdk-:u[jxXZPdQf Δ2Fd"Ȉd>Ae;Sy=8$#|̵1c(4]1d(H!B|]oHWtV?_`wvMDcTr$6\-~;=Kâ/N\bh3!@J墷,نoN{w"ܐC#=mk2=:;n.ۍMbZ=ۅZ0Aj6^u'?&A?j'N&3x?(&%xf>_a6'Jcu3,z9_]T;pࣗ|*o'{ 賂OEYKx @EmtuxL}{'.GbB)vK4'6:@֧$sYylN'6JCZL 郴BLNc0.jg1X#sXoKKK #Exѿ\$u}:gµrM]qvOWxAreIq &EҁI-2{} =+WUknfi%^z6AJ4"PWS5YkAF-,M@l0, 0vTR$d[1 zLvZ=dQ@" ڸƁ6aƼrGq^⤆H(C JXFLk^#(p$w'=tlvXV٪bf T_z2x4h  \!lIN%# Mhü:,O)s!(\Xύň5 TR$2AA1e./s GIxb\>ҨBgfͻM<.GW]KNxFZ?Ĉ,V&b<>vӪ7fe_h}&P\pU"ċõ`\.C=bTmAL&ʌdX:B&bXy^Ϛj.> E`v@k=һ|1@H3 m#s{;rv}||pǨ0Kכ^oRoĥTL$А{fgU3),C-qT9* Y@g%&;^AV?3XOu8̟oP`5>˧ljQ&]]u='nz<lwhfbN1W"^C+wVP]ǴF7^%Nv I.4lppκ y6.;*[In;#ّɆχgC}3|wȘW܏ph=0et.뙊3gyH3Ry] ڈv*𘰙ΘRNLf=͆S=U[B=5pyU֍O -xo#褈;9T8 Gf,H<@J{M%Tڦ$ZB #^rI $it;"mE3&R‰2bt Ts Wkd#˘iT9s >>"ÒN$hJ ~ 9c&.FMŷUn.F9u>ʢi'7qׅ~gt?LJ$]j:0%W}[Q>Ǜeu&*>ܝ>x·7&Nō]'.m/"cfI m懵g8.sNYVmû$-Ͳl~K㿟4OE e˯N%Ԩ5,> f58…VSY)؞4E?];1 ObR7U&N`/'|;$'gf4ޜ4Str_NԳ}Zh8j=!HtBytInf)[8BAT"JAK:9, tdIxޢ<℃fwqXz]:}W->]!e{"BGanY{[gRm]`qfgeI~VN9Z3c5˶3.Q_y~g|3н"7!_9r˫}AX%+&-uxf8qaT Nhis.P,H {@sds[ 0#HY1@t)'A+HD9IK{j}PJ&^HmϞysQ=C_Tq 4{nFЌq^ݦ^k S@HR gsSQĆ9MfkkځrлDvg%|-&2` ܙ!H HCQ4PDpJ ׃&mr'ٯQ /&'#!M'7m-"-?9Q&-ڮQL-II35S6YJZǫC˘io$Tk4x8&EG=sGJU+RA0-kV)%R뼄^{-THƱ`c)U1H*1"lB#U2ar[XL2--;uzk{ǫ䛛Op8:h/b[g @\H0X򗤂̄Y!"KL.k prg#9Y6\*/3r {DD02WtRyb(,GcKIDZXm[0UIHe"/I(`yThB$^ȴb^G I1ލ9o:%XcwuhJT{dIRQ5F}rr(\_Bּ5״^iZ4#ZGѿ9j6`[AJ_tb%bY_Oy2?usŒbzя'!?x;{]>7wo:|Ȼ7qˏU©Ȍ_Ɠ8ߏQ>oϮ&q?6ZI;ynT2)EQJqYJGkY gQmAǛӮYnh>pC?F4.$4tvݺvŴRaEmY:wOk]z՝tCkh?ުdt/PKUNj1Xe"T(jd|vyR͠kGoQk+è#S |)QHI!T[++T :ՔЈX.ZR=Bc`rUX+{(/GMEiM#*.T 8D$:" CEߪ/~H©I@e$gSG'LIEx*Ũp|Bw,p4-EB[NM~C_u| ź[-#EĂj i4t+_N8XP!i4I|5DBh}PP-g*^\˝(0O(edz!)B.$ySh{m#GE{R$af8`8`wp-⛭"y,;d[-[v$K۲wljDQ$XUd=O#a(>Hy)Yr^;%JR302p`Jh؟Sg7qvA^&@ =! ic|3acdX7V PɍK"$P+Dūw@(x>|^*[1v3Jt) R0Pa$ )޽=elcc9[S~;bn't++Ey1ĥ6Ɯ]> PO 0MH GTT0Y؉+gOJ\Ok-:P&NVUEYPE (H'Q/tK8=;9pm66ܰݧ|w~mtl/qt<L?Bf٭n蕼#oJw1,Dh./&}8I>9V44EeՀ@O^I>(/t2@H[!E;0*}ftp^ǥrii_̥3)k_m>]~#Jb@nPXl nVX)eoJ=Ǖ7}M(CL $C"JJ3T24*ВB1%p$ST^r\2V֐"ر 0ZI=6rv` io4K08Oɸ1΋ 44\}j͖*A8*A!X*EhCymDJ& r€ۋvw#nT\m@}0bD:k!y{aq N NTika5ffu|:4@6RijՒ cd戜* g;U\zUs"-U-)A_ϥr,EXsL*2_ 3CJ9d@NKwUNolʩ6hZL/4-KB v"yh2VWnbmrJ4Vt2jN&K0)&aB % ۛ6:j=Pv6?^R*)PHIʔD4u2VҺ~{$8$pJ+yePWmcT p9;6dYzgG=k!'0Ofg-T~P'D"ySp@R~ DdR!Vk1TeP>U;C~ϖ]{t䃨l)z%;F&[).\ tf G*J(ը%+B3;]98AC°Gqƚ8r( yv@o1 z=kiOwMI5%:^~֑_ C^H|/Q _9~|qU! Mx\e9Jr z(DݰS@A5Ftw}w}wwvU FZdEڱR+uJhRH^EKIf1Rim(jOcB-P,a': ="G*?ԛ8V萰Tl|ʼB TspyL>C|rb 7y%m єTq=*^bR-~¼Hmiڬ<ΒR?|!Fm-en f \/!F{!F}>!Vw6zZNޏQթ6MR4Y/1j͋C-UEWi4 )~oKapHU ?;pHO'*jkMd~{tsīs567lnK1rLaEl ŇqK~-Rߵ/M%P7>Ky*̊-fЛԛ ~P Ÿp߭ 9a|Zw1om/4sf9()z zuʊ5(Ş6147kƴaI{ԭtPRvaܰ#iU|5XCrR[EIo2j_Wc~o(P]iyD/,\ 붏#$궓{f>y'].N]"5tYaVۗU-qv=ֻ2yE'/Zn]zՍr-Ӈ7E!]yQ唷I7 !VuW ؔQ})JyLF<%!:6VZ_bbB,GokI㦿'UJ6zE :۫Ce+UV2Pr?~ds{~?Vf5l/DjN_O~{ڮT ꃲ?Ct4V\&9^lO󜅹T;>ſ}~jcQ}f'U~/#ǕΖ>_z/OoYI5Yp]"a3lMQ,w uHB,Rc]$I%MH7qv͇2ql)pFç`t ;šW{b6n޽UF/~s=!mqJ\P V??Ԉ3Ym,ZxpYnخv~m>.,{^izn?hk0nKqq6]9{+O\>o}s,_gjo5r2}鳶B/"GwbM $*6Hs ufo`{]hWy?Bt#7CBY!X$zѢKE%,hoȑβFa!j8'+R"ipcVSA7{u$b s%/_YB'RyϠ.!-KmST;KAtšLtXզjAX}lT=JӼ{*NĞ."jb>R,N\IT(M@(:Py%#ZeվiEjI1lTQB |j+ LE" ԯ"=̕}HD lcP{U|iv9*D[Q31ɰTRwP ?Z1m뼦y,F<)޵6#_cbh~1jz awaV[TJ\_̔,˺٦dRRL&3"x"<(R/Gvk0N'qÃK>]toufln5PS%מM^g&:jLQc㦛紇;<}0q -lc20{u$?`ML;oV=JGw7&E sUjݘ4ڨ&7槚Xr `a%#`%pi0ypIjf i8DtnU0" O{lΠDuSq*<-Aҳ6 KR}M WG:2<6J\:PC֐Ȝ <ļ띈 &R0cb)rz(iN)/Vl8U4^ 4%4NWYNyqRrjϧ;%uYb2f(1`, !Iaa:#(UL)./)kV) R|^{-TH`c)qbDG TiXl8-v&d¶Pt%l}.PТ'=au_ٰW PFl҂(gt_MjϩBW}GB+ ByTZa)ȍ bQ6pq2r=VLQ'AZe2È)Ĩ5t:4hacppzl&ߟ(6 sELRpÍ^`)&n,  wZl</%Euux A:M9!MLvNMư`^ZtD2urP)M]PvAZ`#BiLp訃`Nv1`MJ&CceH9թ2t! ,>NIO4`M8a)^Ib$b0ǬӱұiyۮC~ok LkւRr}0 Q8֋p4a`jBױ<KO`Ī^o{u 'ո?A8`^●SIF憾 #!f爲"rZ#α@.j\\va_9D,0Kn嫟e u ~u KyEo{nMk-{Pdo<8p [|R~^y ˵X֏t+&soyݖ#oerc 2xT 5֙d[oY~Z4*k-HO_?Ah/{>HO mi,=0`U c'1AKZ EJG?|}˽<| ˉ׽}5 xz7]߼7~zObElٽEz6 $s9M' Js(Hg8 G^k .܊q{d+ΗttN:WU< ߆ga)ܼ7r撝Q9<`Φ W`}.h);rPy+A7p{v֐O/ &Xjģe1j$*-:5;֖;;ږ6$#˼KǤ e>oVQ R"htii+vLOL rI !&(ŨP\0*6a9d9޲+6[v`^,uwМ>b=Q-s\m8w*zC%j3:a+QͥM^kq" *"Ap if#"Ny(olmi nMvdPvXP|ۅc("FUsLLoK `E 4Ys-oH.Pe88c􄎅B_ASAo=CM"(.c52Yc"SbI3p3LB3ccIsLD&$S#FkuN_,ƈņ#F,;A9hK>frH_ MxvX7P) AL߬%uN|lAd{dNf!JƂюO#K8Վr-!sx98ǃ3؉:H(: nL3y:BeK+biEkk6gy4|G{f +w2|m}f۽nٻuf)yLtemvMñ:bDŽ3qcBH%b !)L0($$pP1xYEY/CDrځ Xb#VG#$K:DKTx,(U4z88x615,~u6>E ٯ<˷,5ᜈγ<γ<γ<γ<γ<γ<γ<{LP]-b榀əE 8%,"#Fc*1+ҩ#G#@~㫫lWpnfO>_jLgס7r\AnrXC9$:|syټvyr3]9If,zzE,OQalY7ޜsHamdUA;0fVnI[[Zb{7) #qTvzy wk4SN1N1Wzy߽*\17֡>ua@2>yx~| 'pB~0O/j[8W*VY]'heQmLܭЖHN "@:  0v_Z߾HG}_O޵>m$/ yTsrvŧG̘"+zdQ)Ple=p0h{z~]F,8~ h|<4[(h3ҦQY "/)y|i+ y/p꬚Iw=q-_5˧~6Mƿa6{<LyןY%g78m+;#"yueuL=ӡA Iݥt9.O]J( t}g|;ZU9jFhd EsVARH'Zx^ \&4 MA[cH1&Hspmp9!@p4 Frk Φs{нI//V+%d_v9+ĚlfiW>ZZgO|]nqC}ETrM @ L*I 'góIOcV,+C/I (+\{9;,>BX@ULX0S)eV1oI&Jyߩy4B ]:Dn,h&2',RR(lj,YgPObE Z?ރ,m"yXiYrh@|H˨T҆"."ХZ7٥ƻ :EAFNł VE|Pν\jujٰ8i)Lّb "S@23b^Q,$LB^qt8jIg<վY;Yw'vx3xIg>@$cHJPk] EQ:pag! 佴uO.EOE:cmOv,|r.| VL2hFQbN%jAo2F c7N?<|0P<زz3V:co1 z>-5*jn.YOrڊƑ(j11뺵js;a/JmA!@I* {4/x|-)<}O_g/<O>idGv>ӷ?.ݦ+=~ll-n6K4ZjfT#IyGt_Oh fDOЁ(Nߑ N@:w_zgByI&j:O?mܻh~`E-Vgi|2nP={S=:du*Gr"n>c*wd'}ށą _DidEqRptP2QL߃l﴿RqJmVN%w~:(VC瀗C΂4栘F0gάk ^机]ÕJq!Q:eƹ \g@E"Ykclkfd _B|hy#sdG~$~ʡ=k 7}0Ү>򩵾~jo]i&~ځu:Ų`|=3L9) [{`)ދ @ `z(͚^a/lźWj ip{wԴd?j\ۛǃq7D^yϔ0QĨQ2 h wNI˄ ^;i̅ )4Sщ-y*9&.VzyD$7FfKaP)l8 l:,BLPL(W#>W :om}9ܥVt`3@QthH$dQ  QtUbj4Φs>lEw&?Ոchzk.1mSШ@r2B!ym 0E"kfQ{I%m%QKbTO=P!I A %0pUkH&>3]FB2v5;+Gqc #7}Lf^vbֱ3g,)oFL(^6 CY&[o uB\! 8Cj:V-|7 :>9\Z.~DJ3iqhQxsBF9wJdY%O>@`2":pk$t^uOK ) S/9@4XusCTE(}O^LKaEݺ?h&KºS\'>z sXpdDU%0ƃ سC(~_iAjPY9+aՐ! XRs5\V߲ R9&ĸA/eمK փԹYK\(oJ/ K:`9sUūxaʼ^MK:k0~_MqdT]~:cmvl6na/?FߚL͕H8 hʏR?6ӳ6Yq#8'ab\OwR;|qhld#r^WM<`C#jb9b8 SFdVKg}'HKgoEEL|̳kBlбJJHSwHu.?5=\%#U7&UQ~XYGGu3Qa-)4} g5B!"%9-##G3vpBqmAr;)ǃ錌]=?i5C.plZP&2cZr#rt\f Spz297_o EsE.Zأx6N9{Wx~ʃ)u Z}=OVXDI+|$cSZ6#% V\M !~\UfKnx]Mh7=. ^2r?_ToH㐝yFJnT FՌ+&=ل+Uu%#Vt}_wjv4~?ZӔ_YXq_TU.Ħ6gvS^#ׇAMf9__:qEW<ĖUnA]q6@..:6@WS?Jaa£wIA=ǖtϓ?M-N>Ez/8>=+-O}~@P0^_X`y_np.Z Z 1Eв`C`R$a}}S7P%݌f 6A kEKڑ/E&j/J.gð #2 4翮{q?Z7$! {lBd's'pYfq pEnTPEN+gӿJ} qᤆ/iIH MMBRwՏ-|%n=x~Zڍ e.u&E Jm"?UJQ&d'rZuC7jQ~em8mUnX|tnYdD7ånHɉ.8@^ˬR{N\ ) G_U}zwo{i-AHY3y -fb1[ $>k+»ݞ'{ Հcr2E'sMㅋE~`4 jcT_aR w0/^ӂbRqw\:zǥjGq2v.',c"h '^%%b3O-LHSRES(s~ͷí1u x: rX: ,ݥ .yM>:`V )-ӶRZB7:]2z -(ﭷb J#xHc2UFV2w9.N,uL fx*Y6qBP$/|fS"{>v[^H:q2~_=n{N lfOw&L^*\A\.x*} n"ZJRR)+}\녷ᝧ"Iv[UOӔ_䖂^WK^\5m{j44E)eذ䴋\:ƍԠ}(4:)h}-).̥Oc$I%Z$R0gzH_3jy觞``yZ4%[ؤl~#ExȤHJՀ-X̊̌tZH2zP(An`)YQoyxS9'?eEnK_S'#Wnә{!UD# SeѠ[5]7\*B,|FZ d]k^~ogS Df <`D&4FEVuՌn5BrʲƇH.416Q Ef51;"S*$;թyx2F0Cs2>.D= s.3-al=f@^ L䉑:4_x@~ǐ7|ݳ 5Ǻ:4P\~Lu^.Qַ]hpkYT 7mj|ؐGlH낗x ^/xc&7@V 2YCRGjT`5YLNs%JcՆl􁋏..ZMR/L2/};>痻2g{!k『Z@ (}aeAn+Жڟh>5T|gN\~jҍi㶛(iy+SK>~!87-཭U7N5coShAJWWL l3m Tj'=@,=ZbXϚ><_'4Ӭ>׈Lyux]3L*^+]Ww3X%lu|JmfFRBv%$DUHBWt<-ӉEnXkkȈV ZDZh>-ݯv]=زJ#ZkdNGu 3#NUS^"-;ok zQ Zⳡj>63aH^:Jq|C@ESR(=Ҿ}ف_~.y|ⓝضiL2c{lw*G.CzZ4!gCrC?]n±%XtVҿ/5X钠_`DZ2!ن `꽒|et1>i5j?;T÷wV3RӅ` g EM1.Tr;7V>_ߴz~1:o?+4uZ:.vqw.ۓy9 xDul%4iJPbz4š%5A^)U HϿuwuL8qFҞ-9[[mmҨ־A[+ܞ*[v6ul\B4 +0fEM]U_qw7o9ץ|rC}9`r44.KbҨ;v`0XJW$i+{\ne 蓿=qЈE0'42 *xU7ąW6YK˙mӧ%JYh㩁Xc&kDŌ}0#lN& JZ9) W:aPAhƘG)/Kpo Krު%T -ᨔ˻/ f2yn݇pY.VXݙvK#E~T* 3q`#dR$-H 0#5-!%0Zu?ůb^9%C 5Rp$80*g>e LH)V.s6Xk`eJ8u\XY|pE (n?o yLf6ȱ|6!;tN:'hdςAX/* damC|5Hyޱ3jD €ʐDŽYF$Apz䃋D=q"f*WNFNϚc!SS*Gh|#nL83իVqr5}y/]{U+x{ zL+JH;.D>Fc61g$nAL{H0=qtrkc|]lC !7FDTgf^W"~|C5>E{Ma_1\N8ŵbR8\Pol-d9|}xyr>>7q8=J(ᦹ%',_j_ŋ[Ť5Ї"ĕDvMŬbR}s\ أsOmcma}?ڭ^t SW?|x$R"WIK)~#yfJ Pxnjg/Taa&ąW&3G(Iɖ1?s`cJd!{$ =( F~ۍX2&S6vboqvgN]` 5ɀvE_lhpI%(g&ÉᄷRBl Әo3i}Lc|p9 XT))a9kK>U(L~ !7Vj"[a='cGmc5}ߗN^WŇ%Owxz=\s85Ŕ8G5>?EdXFcp(˂i&,!&ןyyBI\7YˠBp@2+ IOc9 -*v5)aJZeA$.KzL2G٣ ]V#gRT7m a}ǟl}c-!iv1-cPڙ-CwVwv.TrIHHP׌ѐȠ?!Tb.O7ӧ+u􋾯]'Ak8}25f#/<>~4}ۙxvo>y¬c5gӟ=9jYO}I C(Q0mEm͏r9ʹeM[$heSZ7ȴo2x1iW~8V}VO 'B%xjMf'7:hFhd |Đ\87RS &BHr(#Tu!5$?4%^D48w9!H2j#9ПPJU5rE<&;,䌑}D,X ٔ~za䅯Ҵ->`rh3?*Fq>Pr|J󜌌*ji#;i=wI~1B[] Y |V \9,e @UX0cѣIIZd)c$)q %bk ɘC-fMX}" de9FNO9k)MfTxY#ʡ,,:1u:0p! ZJ%/ 5EkNQY3Å!}i7K١`5mn=GAp3I4NŜ ڐA9)`ٝL@`ʔ2WL鶽*"5rPUI%V㉝kIgUMގC掸HȐa9#YR*KZ ÍҞ ! ji!ھSSnrFONEȮ'+ 섯OC d^'ay*& ID,^KB9))9&?g`8 +NA:qȮbˤր7葦}d̏Kצ@$d#%8 ՓMf2ox=/Xoh]Ҹ[uU! %n!P^4,^t!K61UvʘE{%=7 ^Ͻj*XjN^ 9%-ΣvU\UIaP _nHB*9&.$+^COrcDIWl5FNS6rB&4kjFJ|k;y"流DWUΫ'aM=?NcҌ[rԉ lt ې83`А $BSHR&`;n>dEuw&"e俧zHєVd"iز9Ùꧪ2"ȶzWKlb />-J>yr QeKtIXH͆se6BnŊkf~Q76s/>\gv9Kl/Np<~/bc *H`e,,@ c,\FɗL(gJ^if%tJj6uV䔝vEA $*"9zpq:kXVq(VV{@wl޲4)  ұ kjSF{Ȍ %495Dm1Ebpn6a_o`|VX[D7X"%*E% KI#A ҤDԆ1QJ%&pZ$JM-bRX2e(IRL1yv0&{ F']-bo7#>uil`KZpzo Jѐ@YlF>3ܻ2=ej:4O(t~|XuJ@R2\v&1}E1@:HxoC]_j zU 9)W%%Ary 0)x'c dE M'/=}>ł&( x<(5e}]KٯH'$ʼn@EEHDE[A ,`xIKg4H(RgK1'o@T !dAZQ6<1t2fUR5YR1װ 㷣j8t2a?n4qf&55|TEo^6^l>]TʷX.| }/}'?KxN]5gpgMg&5'xqT ~`HHm^=4.i(Mx;$rӫ+mb8EpK9Y1w4M?Tv,ӇzR5?3L'<ϋU'!1k'yxcMP NFǣy:ɉ|_N(W<|vgM^ggz::GدdiG88گkq=[^g>hmglS<0Nv]mMnʑHg:^E)$f=VWr4bm Ne|Lv]!lprei(B d-"LxrZdR Qub ݰ^m[ aZLٳxeq0$U'˳5X6阎_Տ957zw} Z$ڑA&(STt>nؾc! 18ha [!<1FJ u Iǐ!SF*@C%s&(s eHnLi.M/ƹF5J7UќϫKg^jF׿]#bwo.^ma2<ـio,lo *vY̟BSہ5w^ojlj0+^C Qw9m={ ?@=A=[0)J@%w4~AHՑPD/ KDXrRnw/ ]շ=k\葌{)}%eё(W!d+<xeR}^A^k[~u-v@Z{BZN{[?>56[T h{BER(`1]7b;kĤŔA$$KT|P:meG`!vHFQt[ Q6{)gpTD:&U!;q`&WF9iU$֍ GOՀѤnaOY"iN!%W֢DbI4Qzo"mD&((5- Ay(ǧ-n \?`` vY܂ofOg7y2;_HMĐ9آCAHsh\6#6E)H"jbԡH=E0Vnjg)&Ƈ^+֠\UZ.QfL)TmyE_o]5 Bb͖mn;+Pv_G;~X/11b2= KQZ7;tZ*BD+33?}M<ΒKU׮J$4)$)"&cLIx6%([]Z.xomVU׉֞FỳE4}E>lmlEn3{n z|ŤWe@ANbF?,:R'M)`H_繾T\_]7 fzo~O\zg8=|Ҁ)j#ыszp%p~>mZnwWi>^;GV2.U=kY'oNxwK ?k `A>v.O6X.&y5-]6>5W6,:=0-.~baD *^Iђc=9[< }+Ηy -hkEv{0hؽm#Vze:o~{֐:cK J@.ǎ=*&#C&QծQծQQs.&Q9 Elde- Y( J,B$¦Qny8H#a^HWl9 uz=׻ >:i"Tc]-J6.::-ҝ1Ƃl$ [E-r.y>h2= J8^ړҳʟόlR:@=A*+NFe;i:QDg j*Cf٭3YYCfqdZ E@@NJ1QE iMwmdQ) Rgh-x;ǚO)[G!D1r ȢhkkuL6MD^_N m,_|mR8~}Uo,pK2Vr:l IeNf|V֛԰++غ۫nA}y<ٱ*`\r9'$ozIZIvEAV6D҃: ڕޱ6%U'rHXϚ zvx@!|PL@Eb?^{433T!3k]OS)_Co T$j&UBldgۯn:=̦Yc:`<1dgR^w b1N;$1![ٺfW%Mf%Mfg%Mj1K!¦A;-%SB&#E TBu)mTv`ig#݌%+> 5R3;.jPŔS춊Ep>>^wpM?M5o ]?{FŸ`GC-rY,rw9cQ]-sMZ z(Y|jZ5r=5=UտÇ GI]ʩ&x5N1hzO\Aϑ:%%O.)MK-2{&58$$w>q K FPD8LضKP" J)C%Y, 1Az\x#(WNy bg 3_V#gbI\M'_nM;K N]0<'s̸ߜ_vv{pKO߮oTy 8p "Y/x bc9K57>I%(EH5)T>iGlO)IC8l'HZx>-UڏC xXw6DA~{DM.t}")[ji?5t{G$&R*,qfD\1&pβa[i xo3),EZcx(t>g5yʢD.>FxK-jTJJyg%$ADEyhG/yn$OD)g/^CBsdad dҀT*Z2zJABEjR5OH[a -6p(`hQE-aM ,"+Jb !tyq.q\,]|kRP\l3@00H2 ^RH.wH3-n ?8tg7y7vc^4@b\$Bbl6>; vywtWᧅ= L%bavqhLOx3ؿY&nNQnh 87lx4kpQP|XV7nP]tr†}fޕE錑3K}H" /%Lc$'N\|皕{(^9*eʠlVHBSqe2q%g?٨ blNJ58(nGg;(".sssU>kY_܇ez7x 'X@T]d/|M7]2훟?OיbHW8<`Nj/e~Fp}_D.SlFLrRf$]wnݲJu{mϛ.M PC"-=$ws66O"e]XDT.?w~A7/xyZ^\;x) ǟt)'dR{{[w~6 nL+RO7wڻ>^y(~UH8q84k8L,>&K8'FVjgLD v fǜ\)16k ' \JU}a۔ % hzR)'CDI 29^סʆx5rLt)9!\MpC 7yO];[x|k^w}ߠ ZWў6f/̼T"cγh?g.3`,cFa@)5Xh *T18u (Z^ TNB )u21 K9ժvZbX)5`|k3#\6*aҡ%.|NZ5:IeYG ĔBdzIkN48}&?2NX6[ܒD$ AHF"V6'Ace(b4NI)&b+Px\̧C\L6F'`܂24@Ӗ=?nsښ$`&QY"R>04p %t=f3W&dT/Hufu6-C wq'Qh'> !LB}>O)޻'֛<,ϙQ6<~?V!|nYqg dzYL? fir\]tZτvzs5t4?.M?N?K멹g v3'Qڻ0q|H7mUQ_v*[!Kq W13ƥr`1WWknzI-a?Jf Q[Šwd?mA]`I9M?1~(]e\mֽ0dDo7zHp-op1C x5 W"ZaC!zwEsVfTgܑW4#&ή%ٵfcݜ}B%7}~+(.v6e]}t8"D2~0tNiM*7 oBJ\fɛpJwX:tvx5lKHFY]) 5ƕQ)<࣊zgU)ϸ{gemǞ;1Ev{(JHT >ODzq҅>}dӫ_Ɠ &X!GF{To~fcH#7ڼ'}VeӽtXrsNb5cmZeL&;C !\i8;_1H{B 2͛/ы2v3k-4bnJa)ȱ=]FZ}c^@:92'ƁgU}4Cٙu6&z;@{u3Chzrkm]G^l=Ҫ8iqzT&G.zafw`{߯~ye`ҌcEw( Ǹ8E?=2 fzr{lH0t MO*/0A:ӆj3$)bfO?j9EV]L $XTxc0tN p:0̋PZRTD.熄"{L:;Ik^ rL*X!6 :kj-X4SB~U(^O-.Ozr=<JN\+]w(4Oq"N2Dڣ0qfhJkf4sHJ~ ՓС6fD] lVII֌٭gх8P]h*B£l\U_w~=nf86/~7hy4Ϳw ϭRљ# ʖ5/dL.$:#wu5t2+ZbAL (jS hїcg|RdLu5v5rvk05dv-8mz#ح6B!d3d!05W䐝V)))B RlmU}RC2䊬hɒIR^ Ĭ#C2[F_Q'cC5m{ X@r!)IbAjZ̛EОt VՈqƌ%8bVl,* 1#3\!F|95sħBTm:z]뒓EA(*`y/>TnKg RgOM$\cȁe%s.U3ad@AK2ɒ'FX2vB26MX(z;$Y3?x ksG+T?h m֗\ʍnBrАEӥۜRVivACn$_|魭@c&YdEM +J02RٻFdW~vA^"oH#̜Hp>i"3#iӶB?"ƀ cCY\rdUD%@("dIb vS3zfq#igB]Fi9xyz˅MtS˥c>:Ifg%嬻k[ nɱGGσ4^=k7U\RQbk;NCJ'1,$h&.b A_l"Հ @WheʽE_q;H oF|X/{vTGBx2Hu.}<1vi}ZacP`r6Rz[qZz\!!),  ':Nڑ3Xț4bIE۔38*"\: 9.qZF+kPa l4xO 1㝉>=>q4L;thQ3<|ނsfV mg7"%* (D.eCymD&((5-9㝡5MlY)kzg%(F37NɅmM~4Eݣ m:MWQ0h~ Ƿp2=5FځGU\w^J?qtq]>X+O+Čκ`3:6k`Ji$H7=6̩vh3h3fĖv{IlZrY8 Vi#TFqKUv6F* &(Zo H[XH2>Y'tw&zl_J]w|:m>vL?* bBGZQXxac׋?*zXd_< X;&W499mC>s~O> kCa0FM}v+Y/9mJ~`.,e)1eQ΅ur55>`Wy6̴ⅪΆϪ~_^o6SӰ%5H=#$oz/6D8b!׫*6W_S*SCJI&1yL#jfyޱ 'i 2a.vV҂ۦ]9_e '(vLIL/8^azSլbMȭ3zK%D Rgh/JnvX\ 0_Cwt =!P>*m.AtP81 0!BAT֪],RNa&[Ūi-}̄Eg4cj`"$}$bcۙ8J7Kg3Lv}oֲ;tI ~P=Mqe O_.!p e\J@4Q1)HS,)m(4 ~&P"GעvBR&YQb.Itٙ8-x)z0pRV5[K`ZSLV}dGKoU{x9o;*S_bTc"JK |xn+o}iܟNX+޾ 'X/3~r38,9tauώ+J6j `-9Oʺ3p&{֮m ?^ݏWܙKBIʰW"e%, : YU rIN3ﯜpDڤ_0iӝ r''|ĹC2jyFEL!M0c^KK $uycεkz,mk#5K9KkY٤]6#IȆ@Vj;!CqJ^86^jCeX{vck 2XO'$#J*B&z(\JH`xI?w~ܝ(8EP哷P *`E (@e`;7n iU#;+1Q0g! At8R?yFse̲nKȘٿ678t_rr2qMek4N~\U_o_6/; hnl׍?܊_5/PK y|6:ߝ0cjS<焠oO/sGp>3Bj|`8Uȿ.m@ӄw6KR&NkWi&S8~u!5uLKB#b=[,DLI/0ߏ6ON_VY٢?Ƈ+dKReW_p4'x:wt93gk) -mzE8Zg$88VKp,,MiU}G={g>`VޱUv ~AϝmWj1v" :xRe1R&s (* c { 4|ۥ0#l:iFfE"32y3iKR&aS4D5)K3͝d4tzpӕN܈~ĽcXEKM67x|keA%C_,l 8V})p*UJ8GɈAl ەip,0!]h^ج3I: wΕDZCNnHR^FF{Lw}Pa 1]ݶER{}dLA7޼=#E4cǵ@dDK)!QɁR ![%HL.cz+nΤC]  :'i=r:փ6nGYOSOu<OL#;q[i[ncN]{~O";"$(/|2Rǜ0 H"rH]]4d뢲ȫ*C;Nׅ&ߟ愛 h0%%-Dh9! mN*R7d0ڽh'UҲQw#oP%"ۍSe m49:'|(3q֛(BfS(vܕBQj7!jtD>.ۂTo=ՋY''my@\X}ٱ5Wkj/ؒO ^y}OhK W2'Ɏ)MIEH 'Lv8zF]ɹ(j!cTz(U~K*#e_ l8ONkؙ8=(Y|agX_h;O ձM>մr[uƗ4kOOhtdQw`o;^^M`w~auE)*> oufvwĤ`#hb* Aȹ:)|ϤY7Cs;Du:~q d3+J#TR*L$c iecۖAhCaM%QBF" 6B(KQ2ķ3 ֨ RVkޮĹB$]hrQ$jSw9X 2H_Pe21# d Nm啳m|AzK!K/_tBp.x2(V%e* lR^F^u4P.AQz#l"v20~KD+ b 2 (`:hM)mWmaTUXQFg/`dR$X6C66TY{5o<ԧ'h#kd}U\7c;ƘI颷CC } Lco*wn[erq)̅C$J"? $<e5O=j FA!LaО3YvH;6OˡL# [XY.I7Bm֎ԭ2W&(+'=:RVZ ]f1,gJu쯼SE<$8`ѫ׽'KeMV]%}}9)Y9D"s\YC .p^I-Kh.z@@BH# $}  ǡ8'O!1.j@Sn_')e) ;A*Z4=3b.(83Zco1ӬDZ|3 1Gkٳ?~=|_@!(,lh4_lk999f.2BEUs6)VR&x˵K187iY\bU!Հ-49o]rzű%!{`p#G(3vڎy&_/e]s{wC )ăgtA),yJy+i}:#)j#*jԝ_G\IIa"k jgr(48EmbS*h Bc-hc,'C\ƖwAU&DB:F/ijo͜Q"(Bv>͆xl÷ν>e~ K; ;ALsqh3 , =m`pCV8jI5N3gqfΤL"xs8$6W8HpJD9BV*`V֔!"liJ1#1ّժdgjo^gcB2t9]y,8yywbI_˲l~YCodžWa=j U_u b&>xCPo\!'?\nQ)r!R yƛcK.nv]:RJuu>aX8j*#L vu/֭j \JN3"@ d:n(q9 w7r]Xc/v9:;oqg/l~k~ĤBl(MI=$PZm)9W/A: Z_2Ĉ% ޗ&[cL#bbāFa5<69#,1פzw&f~~g}CyݦtڱKryN+&fNXj`<ڧf.S}j0'ߜj>a``#L'gʝ"o]78̀v`P#;.MSgmyF(.'TZmZkpoLY(Q>qc y%Z%q/&x~Iq+х"6P,!Rbh1";oOhѶ:&WEU֗Z&m(BuDz;E\,[^ڢD=ƚZ6;I:"P/WՇ]JVʊ68yn7Vbwσ'?;kZA6e蕪99" 3y(68O`48"*vK>j畮I% 6 5s Ou3@?[Øˮ*o"2@.r 1 RSxl3ƚwP\ (=% )ּ('b4 PїhU18ydV9HrIO:bpwklREKPIJb-EYmg+\.+%A0S8feacsG]s:I^d] SU$1 䙽ild'x0?ؐnnfK;{~Mﳶ&<H !j/ mW-]n0ƽSnb;i6T2Nsssmǿf?Vi#xSm9.Q^l}][rlKyb5^^I4[Dn!A M"G=*.ՎT'Rq`{,q`{lq`{4q(@;H Q+w/FZAJk.de"Zs3ϫ4 m#|a&3vOK3|Gס]^z9n p 5֘)jh"JQC؋~j:\wvzxbIy~ځU!SQC@:q70#0+xMϵw7޶[wt#.`+*`ؑb5qWs&v[h,@qywX+_/(N{շ?;Tx=xZ(^{"[r&œ CѧhsQ P&MM"Bc+Gg 1[*6ȴJqF>cN%::6b-x+s>Z}r*Lút9N gKz\m;?f7_ힿ"vȈݷY5 b\V8N$e SxA+DW,F7b5'z)|ȣ$Ր}I:M8W8A&Xl41%x?0EF9]ye3mq!Ȍ$mw'$E5+ )+Łk2ƱA^+Ny,̧"rI <̦N^mONYC->{(g?EEo l?0^ě=_}7km:;18Lt٨OGvu ټ\|6 Mw@n&h`nwv=scI`gfj/ZaۺTnzLhs3s[Fl9Mlq.~/Z/\ϝ`p4#z __RZ/^xi2m NRUZ DVL"J:JJ*Rgϋg#;t\^Oxƭ`| f+F;a2%&UK˶!gt1y>dOk30/è͍Ʌ H)*f\lFԌnj$G1i ͜kPx@ }0LF ֗¥8[ R9q$&Ђ13 f -uF9~a?B'6ts1r~k[ ]?{v-js{6ȺZfz ~[]'F4myݬn{%_bIa݇26zXxso3mw};o=ZC[_2z0/_sU?_o|ܻ;*N㯇*Nz՟M=i}~h&n>ܜzݹy^HSνlڝs{wg~Б8bX%s=#|co(+I*V:t'T%\ ʊb6_$,6\^?t0kx@%KT*:-~#ZL#atFDiX8EoH+PC!+o,IyJs0٥^CBlWZL5e,b(B663Ō6ǤcdGV Wn"rc"22<]sՍy?uNOB ͬĬ{0받Y?0HL<'ǫOj ϶|XrXG j 1!\Rvv q񁼟]@m'8/e"XX[d-J/hV5QϨ h)SPЖge$U)3Jx}kjH{2"ETͥ,$: ΂ Kwv3g?LgOE;ŸD݃ebQRr R >pdAJ6 ̭yt:V(\LmYk zykhb%cxіBqZIei9A&OyvmB. W%L-]ggt 5T*A}6oPAz\*15MDŀV§%̕t,+ xd:и|_-&0Nf8ŇR5֠z.*5D Ѳ TEI86Jj{_~?I}mhm {nہoe< 6DL-Y1 ~ڞmAqC,aT-~KR_qyl&VL RMƠ՘eLo ل?{WJJܗi]7+JWJ{/!pΑoB$fC;vT2w|4֖*M^ͧ6ŭ5Og7oP .V\K3T܊+ݕlŕ*<˕ ~+lq \ԯ՘ HuڬsQn2'͗ܙ.&L`~e˗t4ۤ(?#[.c '(N:TiKugX[ ҅sܥF *[eY -%?&z<@Ƀ> nCFa@#%=˜3&f=Fo3/PxJ!Ѓ9Sn< Q[@V "9T`Ƒm+պ)kBKK!TG+w} CԡIY!<,bEm~j&+CxrHÂwYV{,+qe(̲OY&ʖ}88 '1J9rYI <\)2r =ܗ?ypa<096bs cp!gw6bԵH~2P%p֤)D ^D}>oBh~q%ztJG?N֝CņL WxA|@,Ϗrʢ$jv>xN eNϭ<:̤.kBm Y ZYQ{ͭ휂ipg%UCTڠcy9/<`$st2\KXvT QDFaID&%mVx ZwvԳN~ٛ h8>{HPTi"awAsU1&k>FAP/as5UGNekb1X8RJlhMn輤9wNx=ԪtIeu˞3aC2J˳%À,Rk Lr;&_sL !pCn:A-ޣB3ԢzD%EF%TI2&Ń򜳬r ЊZ"- "zS7=[MOoy>aIQ+>JXf`\tA ZP14҃ߐ-ϞuEV3iqƖkOs F-BŴ/Nؾx{6(>9lR1>Gѫ{ eJRژh]DV-w6"!ERM6:^'؟Wj".7U)wەr;ߖ£{pfoe,ǧͯ&%d~j&ilTbwQ"m-߯Y65_t"i٣TGA?5n L/f[IZ@9?nsGОǁ#A"1~&] }W ^ӦvEGeg՛;0΂:BfTvEJ,MmK9]'?w4*grBzec.{'ܿcҨxɆ'Ma3 WONiw(N;7^-2})rG<↜G@9b_[d.dJQv~m(Bq.0K*/u`ڤ5Bewwz\~+V"Ppu]$%r1<.|庅kZؓx:I\ݻ<4`h]"T/}6gw3=/{Q'+;A㪗DD9A"Ook)T.xqFsQeE~{{'ʴ49YZg.R3ضٟAж.OnET JJ(}59zi+%[Q]c9>zn=9 pUDbpI<`VX4:b*xRa2MvJş6|XUjOW7qS$ JX20-uO,5#B/m{/Y$x8޶YH 9Sc'o}xygs!4usZI]1HEDgyo#Tv2!+ <q+X{jw epЛp2rT0Ԕ ==OfJyKkB ЅYt Go:oG[0MV{AxH޺t6/tar/صcu d%1g}}?*-,۵u闘q)+q&*(PydknmVȧ{5@]lw~pH:jٕIX/>l>5g3pSr=(w_`6:`iٌg -N~GE?TwnCĎ3XRMFoYy_*ջK𾞂_eD$cA,$1;7!$+8'qkvolG|Pfo`A=_YZC#$P!֖!sQXGpҖ2D(EJCV& t=S5<~/5\;m_3jV.QtaZ1V,|љ_mboqݫ^y2WHÄw~`}gov ־O plV'<|{}۫(S|_]Z~2ugj܏B٭Oyo[2gT{m啐}1y<9[S>zrNQK0 RRpEBIyef֚v)& V@ wyM@F9RD6IGcL/Td&ЛljBbfF i(Jl[ʨHR֑)`ټ6TKzjqL&Q7)4kbg_vt~q9ty` feXbG*/۟urolg3}=ޑOlKT>|&VH/S@H2љT%NC\6"ȶh+)E J3''o\*!9R4I0HFfF,Qf mc,4#>xIb ˏucG>R9傖_5p:=c:Y,;bc *Hl:$&b_VYӅBʾrm<P?LJ@ua#[Sr+ !` "ƈLÈgi˹ jҎQ{d{}6BXGEo{Ef`J1(X9kXiwfa1!EojbPj cIYUYyLs=/0b vCl9S)Pq0>}++"ƈFDq[tiAX EYS B-"F)0J:l*@P:#6 2$$U8S$ff>1ƈLÈxmא:iqq;q\#.8hȺ;x<0>WYz=*7V]gއ5Ʈ>NZl ckL$S%ml; :PŬ#2ֽQ>3Bŝdyu/ GA`UP@[ȯ52fhE(`sp/%j[e1J9u`ɢQ Z/5!%3M2S2),HHdhLb?U⭒g: +KvLyG uadq>)M͟YU:܎Yͼzv<)$k@ %p^脨A)b:%)kl̆&%#!7(N&Yz]cPEZ1 %Rx-|9;om<hlg9Άu/ECl(Hk ^IS|)e״C"Y±ZG( 4OE@}N?3CC5pH2ڈr%O[~FQEFՌjr:It-`Rhd ZsȐ$.%gRhؗKB@h8ǰ҉g}7fGTd`- Zlz6U:cTNcI Gkk"²2!y=s-ToS`z4$laO,N6|FW{vYK|~,Z`<ؠN6ӐZnl%.~\WvpomnAH8M@lH&VoPmcAx2{qSɠ]v_`m%ˠ)x:XES(B듀PJjH` `N]q!-mjm'eSQ1:V,Z]o%a: f7*J[{݂5by+݂ݓ;y3¿{D`ff7O8!C?ͺ-&*?o7 = t̝,: s/#y~ap!Zvc?|:}R"CL/F ),SR5y!&'"?UƑ*b lI+Hǔ@rE`te#f)x 8>nߠx?9esov 3~?IӷKC]ixS*Q4RTYxǍ!7@#"<"T]ځ^0Ȩ xV{VD9Zj(as6zbJqdo R.dWA( TlwQ(fHʚsf1nF~ώ]Z|fge+vpTk-_fvSImxfˠ"6 IJQJQ&g͉`Pk1@x|.гzFyn#taC!Q}&!D"x+c*|AiP4hl7"!7"uikek` Oc&JS=fjV[BH1+$@\ V?/FX;lXт8T2tH[֞$񿅐R nt9gʗL>KR) $ia2gKRBi !"ӷ)`}36r?$ȴDu,WW]M}ymSq6Y.A9G{[OہD{$ZO[f}~tɧ XuanuXdlۤwc[įW0=!컸%[t тr=rLvUѸw'e6+S E_[TLBRhaAJęr 8剋k緯hq `&#AtM&ȑB)/R;Sm$g}M0lTκAjw-ORr-/ Np1K,:逪\"/EK=%w T߉&W,T+T־)#:W=ET躗Զe߳5E}ȹc ,R9f H(Ib56f9º ۀz5mk}Vʺft]Bm].k%5T*ۿ)Y믭U;=m>ͯv?7n۹y~}osZY9.Q8ap>ZޭޓJ57izlZ[]?=']F|sk{JY1Y`ԓɊYTbVizV*qcV ai:6"]PСN';,vVҎmJY7Vn&߃.I1e;"H@@`@A5E )ZC00lSm+㜬u  P=`MYPK,֭KKȕ+4R4xhNw A[mT & +Iyt7*c.eZY%^[<.'0?~{?5-@d[LfuM*oU_f?]6?w;Xۓד.w?ۿ+.RX}_$SuXJw'8K'e p M:5)\{koE{o̼l;+_(Mx{ү׎6qG1~LjQG9Ե<9_%3czs "*,.0/O~Y}])['yz! KDxs&8=Yy# iL˫_rlyx^W =ҳW?~FBدE^b?Œ'4tkWoEF 5Q" #fl4fqJt9_תkE<[J!۫9&6"쵿i 'EoStQ5pz35h+X}ޥl֙%Vm#ٱEv} ,6sqve{"ˎI~edEv$˖[$"Y,a]J->[R!?=ҧ>9*ud0H~>jf7'zwPd` 7(PH"4eTL#^zIڏy[zv󡵊MzwRJ]5$ _ɅSF<`cȐ"i!i ׂq *(y8_{Z4̭4I F)W[j ]bwN~66}]ߎ`*Z{I~:@hcwgn?Y?'B#>5bͽlt_`ƅJ/!ęlh:m-{ C=ASQ{8rgqpR^UЫܫUo agiir͗T%:2 E4Z0M\ʧ%Fh"-u!iu q)Ɉ2 HHEGrCVxUm&JIaKKרOv;}$.Mˁɶ zr;&su?;vx9cPA:",@ c@(L.K&U3%I߭^P?JVmj뒭);EA $*"9߱L;^Rpy0^+8m{#Gc6BX9FEo9zEF`J`PZb@kYXPkmIuS_F8NL"YEצb1E⠚b3q66W]Ƕu]{[t@$,%E)hJBQF)mt"X*>i(uBXpĒ-CN@gbsI W4JKªIHuT3q6{ ⡎'w})ٙlR;_|L 8~* 'm  ƈ1XpQC L()QI_܅_<ؙv?_p.fXG ~dm_!THKݴW*/x./mBjrnd^}p3V s * )&F4٫$c ( G|% AhCaM&QBF" 6~B(KQ2į3 ֨ RVcۙ8OX֟3\/sWo'" iSw6 cU Q& fD9ZLA٩$?BRR*_(E)TBxpE(-;(VQWnSͤ+CNoM$8S5Ti:, }D, (#πa jהrk:v ofXɓOH9( R*uoagR|2T<{#% vǺ1$#}nD>hkA:ԧhX`hhp4͡i^Yo6/ 6>1>Bj9.(M@aHh7I^Bď:67zŋlv{%&DI_(g1aU|!IZQE @]qVÝCt0R]ՑW > +^ mjؿ/Ӥ"{q2y#B9N}0`.w1!% ay(|xu}D~T h7@6V 7 &K"?Z|ڙ|j|H( $KT|P:䎜BDQ]CY:(ڦQY0 2d2V֠"8Բh Ĥ:w&L>g: lWG7|lybj󨍉y6]ٚWm1ϵ/MqX`$Q"]2ZMQeƢo';6շӴ+\kOPY;Zb%fV>e:Vtw&&Ԛ!fjW]4:6k`i&I4Hj2Ɍ8ꀶ,RnC/ WkP. *-@CB`(I3VUnfkc`)bL%"0I$ TT(gХkؙ81"{枇IbzLkԔk#.}+سއz5=| bALg>z\tu6B?}xdƦnn|1TVggǙpf8 g2R/9PkIMri _gQ46Bd Fn. sGI2fiRHSEF+Q̀`rL[(A5eWIJJO 6ͅ!squbL4RXڥF |q"TJZP4CjtO0<^tE+(U>4*F- ؠ DZQ9RU`q\YԄ6;PJ;?^цYYK{rgIQMj9M׷my-bPOgr_tg:Ú<E~oMNfMVnp6ȗ8LgDcDijMpuU/C xPEj^oR'sWn|=#Nn9ǵZt ]דɔ'f~_o!(1 AIKyZ-WV._E^?g4J!`z[o7W}vhh?gg)t8z))҃s+[}?O^噪U=5M->[R!?b oXa+>~b\04#N R[r?]orJ_x,,}i2~gmp4g7a\ z`< 6bnr7j!A" I6ʓBbh@MkԪ,9o1b@>&Jno9WXì~ݖfd\&P$2X˃Xd#ٿ.fdS`0`μˢQ;햜[dGv,˱(ʹA'vJ6y/xY<[AsTr69M ]Վw>^cQO>V|󇫅M,eex:~]0 T^PV݊Ձa*޼e.71J7(xG%^럿z"Jp-Io~~C+/,O@vsCty =5_eY3ǖ''3.`l}ɰ=9HRβ *c%T25bՆU)ѳ.8$I'~n}VS} R1tv):VFMήRFb\mem(j)֣?P0iRd+&o@PdK^ܹ[sh} Ɔ:0YS2BnYwAȻx._Ͼ, m8"Iۼ9]b8mCk[~R[ n_ȳ#H!S:4Hd-W9]I!W:{= }3 KT[w!cFRO%0#Zv´EbIR3clόn#㹹u΅vʅ{ͅ-j}f7}//.`Ŝ$7/?rR4Q{Dr1&V7e(BPjR o97cCF Q1 R2`Zkȩx#c{*D6+Fǹ>6`՜;]^5r9kt%k)kOQΦHPpWf9"E8X  @`OVאSM[ 2(,:CkJUٙjBvkSP ۢ꩷ylχ Rfgl4xnF32a6U&(UYC&ƀ5ګ6kkR q"k( 6YƸ~׎Gh1EGB!`P&aI-Lr>Xa1|y?n l^-')~|.>iHQ '$<ldgZs9Z.Ed'2;c^řutd1tl57&G{ ins0NH ՝WOPE=€`H`hpZ/hvRp&8zlU[k5%k8!%﫷5d&i .*4@ &j%h8KX Iڶ=0L)Mݚ}IrUiْ ra{始ȸM D߽áW{8 5+쉓 舅R}ƹPy[rR)x 5Uod߲gJZ3b0ooUNOGmG^ OرVfN??MU;荠'y"տ?]LRc I(\뫽ȕ5с]#۴u.Lim)z>&e9P˽VV/ZQ̏õ}a2!*# E7PvsԵT0/"3rѢg7#:nF!ZjHɄ._ (U)JrX+nz̓qVK"բjX"vm6*c!I_W5gdN~! X:ˡI6 +]1.guZ|;bVmߪ^~~x)D5*T\L9$&9 ~~F3 `}f)s2Uń%E|_bӠIPC_la!VRP|` .&4TrY"V3֜'i%//aYܯfgVۋn?*p{ >^[gm/GTΆRԀQ"y[k)VN뢅2Zk)a mU*k9$ԵF-6 HF)PTuж+iIPqV5g{?^]^^&k'_{|aNt+&fy$fc8)`Nfzx*֭}ߺr}mڷiT0Ni}*S7  \X8 *OzώnGÌ6;%(gTX2ef0Ҕ5WPud zi K\Ch\҆jHQh7bd1!9oh5g;m9tL/g\7n8"iGYZLJۤCzg4W>*'C1y]Ykl552MJEgۮ1A|@R\ D K^WUlJQ_*RL y:OyeM4zIr6lVKUf[sg:0sl˷ T<n'*4e-J(S!M, ; !<@R*Q?!O{Kf<ŤHɑ@ qP08y %+Ϝ$ē&t@-_B̂ CfQsĀ́\-6u5Xtfe+Y$LEZs$d]#1攍"z+"Ҷ:4Uy_!{JP j4T^Yi鳇&o,ペM  Xynt`qM_eXeX图:?-D+)HjS9f$꺕:Rg_wP94$JfP,&jMg:i\J{(J{hJ{0$P5@jiNVy*c4kBQ Oٻftgc0yE=b2>+P׵zÚ<< NHЧɋŰnfu0W5T~zHrr]u.(7 x)j>{K C9}l=69 PŢ- b!i{ȾrsCEv,Kv8Z5gR<9].,aV#mL.Aw}Q{%vpǻYK챈_c.^/#q1,KlSv\R'3[ [p9UZ:P6s}LzBOG y)~Ӥ+οcZ9Wj3|]-'iFG,ZO;f}ms+85ѫA'6ūƞnuؽ#tmqwc7_ ̗]T0/|цppœ}-ܡ#Ո]U]'czy QfAJ("@=9`)$H-(R\w߆'G~uFɽ ox]TE f2]J/9R(E1%U HqJ$lT.QWi:~WiO"TC^8zz/z=<˜x_K5ȼ ky֪˼T+uyy2/^z?io^_S qpa0""~8/N ʂBLh/tHb 8AiRJ w>}ΘE }ȹ1ؤ,)"uIKĠ1XƁx3sjjNnN~Q!kțzXn0}Z޻͒{c6n9cNJo̶+q>z㵢uNٹG=n]ssǖܟ;ve}y(#Wމz*RN)k`]Kv .-%X}u .%Xk`]K8Kv .%Xkm]Kv .%Xk`MlUKv .%Xk`]MPzNWsdkk9_unG 7xd٢+F]1+Fw6[]1+Fw`]1#@W]1+Fw<]1+FwbtW]1+FwbtW]1+Fw]1+FwbtW]1?=".Opu .%Xk`]Kv .%Xk`}d C* C A]lBДs0$@x(S1~MAˇ]"嫌!_N`~~yr]u a22̯.f5'5[]m v{fqJU^O,? Ðgi<_}KFf,M>/˄jcv4镮E ϫnΨ~ 0{&vAo/:e^dm-{]0jct5MӬE^L?/b=^Hj ?6/&_l3Kqw#7y|ovn@h& L;~HeZl46C7C eyќI`p:gVQ38r6_$,Oð~o Nd4!J$!r,֓6[sWڕ~(gSH~0t.y\Nowst^=p{ޗ_;f67W~Ju3UKy_3~Z\ֵ(Zrw;;^U? ;,¯b L=?Mk. y,|oo~gQeZՕy?n?y^[ޱ}{ޟ #>J<)qZN<])2@L͘iZi|Iт23{U:qkmͰ=SBeEb!$kS̔286~%DsDQ5ƪ9=W/{릧e>YT8 0z7@Oj6 E.l2%(,M 0DPU>C?=׵9.=Y|hc=G)5S* k6@-.L wjd u ɤ\W& q CZ/e>JuP?< 5-Mi8]Xݟ`oyAs@p4r0i@VH*R͊ertˤI5eȺo!{(O`ؔ{𠣍B[^uŲ==BfRm$ S:fQXe-D2љBDq.R!lTPmilxWu_t<Jl h[ʨHR6 [EgRNj3s|Qèᖣ|{8 h7#)ݼjbOx%ё]@hkD8/xBΤ"PK/ɣ+l|I(V(PO*L@`}*!9A)}NN:5XZ#c3s#c; ys5B۱"rOQ`Xyd9-7`jcPAzc3!1+h|qZ@($P %eU$El  5äQiUaS;O2W8Q W(A#v3s#vs4;Ƃڮvgls8zEf`JZ`0(-59IJǂZ3m,}S<,y (t5) r*+**c*6 :{x̜x 7pGӷ"o#bG<ޡW(2)eN`VI/RE#)"B8tek y7T)p DeYUUɧƈ̜È2󣎌pØrͼ䙸[ .BŎy|aOҕ31b 9\ (ÃFZEut\<.yHpߓPU,T*/m뛵۬OYt4j _ `1D*n09Ѐ&!T ]_Yʚɚ0. EBJ|G"Gwd%߃ل< 6eJF ^{LC4lQķ3 0ΪsR1nfs7+2ꍗE1 鈜5@>Vb1:*i䃉 Q`Dl:V޸ɷZjwc(3JAhJ AC=kt>P+eC:pV*l @? rBlJ?g+14I|X(FR+&s @Ѣ03)ZfpT;{,A}8Ӂ/[έ|E`+իP}i-Yk7T+}`dE6L qhbQZh1e0#3lT9u|w-]|/TPO&Q({~5E@-'- 9(3Ԫ#dPkcI"$#Ibm  s8H#*K dhȫO7\?xڮ1^|X1[On*$,·u |{'̽P0g~ޕ6#"ngd,f a=h/[Sdױ~)ɖ%vܥ 3"A6tT$(`4Mˡ8^pZj8o* {PV\O7ֆ7OϾLYILF&ƃvJT!C $< OgICLgd|O-$ &T1w>Yj]DVĺqZX,7JјOr|urQ,烢[T%1?"@,Fw%Ơ2N7eiO I#y]It^uWb认S I% ^X .&2N `JB sӢԪ[ʅȲR$K,2JVA=wڐ \S J<3%+qG)'S{9}soBWK]ѻsɝSԿj\zŧbm6B;qDYZ*;hlV535=|$:3mptd)^2eF 6a!%0Z:H)ꖅ(\4By>r\69';^`͋ ?rAcV Z7~93Uu[G$5nϾcr,4LVwdbҽGwNG;N9w!KaViM,ޔ·Ed3eC.Un!'^K/oL1ahHA֮pZ -$n#d"f*[96ch5dJt6\qwb|X"!6x8[ȥ}J dZTyA:RJRcHh53DL{la~zt:9εs1D he϶'$ʁnY|/z1dJ^={gwu!ߦX%nWq699 ^#8ˆZ\aJzGtf6.Zת`t[ÖwGuVЇE7@r\<.T$˩!Jf%ɀn]>a'(oŜ&bah@.??oۯk`~ 6Lߥ` 6q\PIȕSB+48*hdJUBJIWe~s'Ѿsm6L6ay()˹XȕB<(!* =(y Ӯѹ>:G\s}tѹ>:G\;ڲԒoteY6ce ,EA} tjLCmdzhѧ:gQYTEUxj7(#/hFh,0L;&cQ, R;"yɀ,Q !5$?9^D48,Y aDٌWK?6DF~Xth>*}|洵F{PGHw4Ql9 >yNctNZϝ7<$ F.R{ɭ&ͪ^NOKTJ c9=z)cx/m="D,5)yW+|PY`%|d"xQd,&X- Mflt'tqZp {aEcMLuE:qȱ!+K|NЋy9XNp|D4i4!mr.ݚNHM)Xaܣ0.JּK5r4VgPQ(ӍE2u`+NuFMn4Ѵ'wt&J1j`dF0gά a,|ȧx7m?{LYyTx(9BAz&t&t.x#wvu ]pq6GJ> n߷sooۈFf5=x^&.q@;DQ82Ը!2 S)+" 'b?1+k p=`zv{Uh:xwKaPzc`e +1:.R.S(M*Eb`VO BE#^Ztޒb^F咲!% ǐB/<'hoIlU̠CҤ"9'h+gV(H<7+IDF1vU#n>^ܯwd_ ɇ2&egku=Mf;5ܲ%v͆1_;"IRa3!0fI=,k&Ch9N`jQB ⩣x&YUVkצy$(d(-sVs3=(U\)॑1ʃHS:QL P'Ϭ9^ˑI"ch1 e4iq.哥L*rR8r/դڱ a -p(`hQ?j[vC@Xj("A_VWZ1W=:)SqȍT}~`[B3ǿ72d{ FZ f)RXLotZ 㒿_W{OlAMl+aT48Ofi6?!+~{1V'/^֒"~.SgrHĝg}_3ɭdfӓ0'Xa~/0+3\RJ57ܔ4; ḩRݷ̏qxNMbALێ Yw-g8ew(H#]qXX6lVzh =[A#gׇn} f,;mo4]6#{gb7J+ےy>d/Q`' ۲#ljl?6_|rLs1.LSҫWv]1z4l<315e_Z |dxX/ n٠ը}fd4&&/)9o;hm׮ya-D0{7Ʊnέ-<<=}%)I%+e4RRri>]:Ȱ>^G T jqT^(a*ܰ疡T߽F_~kÚ^Gs.>. rl4H,a`"XE5v!N4v5YC!HXe f.&d%&_R!j5I S* "w)c@ r=:@|+#jxMn_CSb n]9Ua%WLǣҷ.u9 d]{^‡O=.9V{I2PL&2Oʬ}et]1sXanމ1iwm׈vgJo a%1dh5WEX2W^z@Ӈ[_,uR;t7,6&\ctu|-_E JtKZd5Vh|FK!6};j=Ywأhyt# 2/ӽu!sXd32CBth ȾWR]A8 8Ѣ*阄rjY~-r4iFm99=Zn̈tXf`q&6\r*Xt΂u!ty!%'>wN4at1/Voʬkwt ”~ FAFdcD*8k87$ IU$H?l ~T["=MBp˒]փӚtgIR[]#Ժ;nn5ڙJ{jEhc3i^~U14Y@ zEVF~Eddq f\*Ie21ZA 0K|m. Fdh+欄 VC/8ς;Y.11K9ha u;T\d\e ldJРˏ҉|F'SXt>0G(uFJmg&`9kql6)uV8Yw,r=;<'υdveⶎ {q|lbcŤK4 2QfΉ )y DZ:J$Y7?>i=Ҹ$r=Χ'|0˸a*䧓Ijv)W@ؕQZW;Q\l2cu:.N{9HjrFiea^x`8}~Ee|Vx5B^dTrK^q=adז`8c?:ߝi;M#W/-Z,ccbN{e36]/č"ç*}*(F)'iaE >`3Q4g4~u߿IYOZw{Ӱ\kR{3dyO矋s74zM}G]%9ƓQL)Q3QOՀu66nYD[=ݥ<.ڲPe_~ IQ^ujIeqOJܔ~ĭd4IFFF D͈IRHL^Y%b("2c,[e ĭs hIlf|W'^2Eⱚ:1ȃ=%dњSUJņN,ޕNq7o}+.QC, 8fib [.Yo[;Ka2m2YA΄;;m@9NIF|U:x2=J=|w8;TwvNH8)[K4.f?IdM$RUqLYg&[M(LI>at8rls鈎WWIGVskE̱]kZj~LU{erxeG-V'af~:Z%wR} ]kWvŌܭHsVL'_b{x׃&3v6e>[~Uo]\}A1Q*' y8sWY(ov/=0-ݻFzHJD?l_Mz}]ؖ4la"*-4":wb*ޙ+2Zpk%FѶD\g֮/Su1(6OU҃i\֋gV嫋`]{2m:>OL"u~Ѝw\v/m65X +qss3*G{Ђ blt_+^3h[UJ5g2}!TYēQxczŇaRfE ?]KOӅˬR];. J0R Uѻˊj9^GrhsqYd@ЧY"B::幉I~ z j ^"Ѵwkm]g^Lz!eĹ~0< -ٞjXӡvy C8bY]e@w)kIы@2-l2D&yլU~iaRnv> loOq4g׎phEƤeS~4W$)`(c9&Q9p-bK+3k(FBR sF|l;fJRf؎3qnGl?57ݕvfcH,4lTW$& :!9f %R@gh+7aV`Vt!L&&H–" }*@Fu2ag܎QT76}WPD4#q@ě4^9,`OD-dKE@ !9FB "1m@r2>X20lɪ*pe+ALjؙ8#D] gL}uv%pqś4>9f$g,6w\.;ӎ~'l2*pS2я[TSH"\/n+hgA1j!P#XG3 |P_\O['N+\Ou@?+&\` ÑJ L`MwdAu]D,0#:8d:i(rcJ2N,h%UĹâ"dt{]'uZB&cmF%`0z]*ӱk|KAңR҃R*1d XYM鹖"Lִ!ehj9:aWuZtH >JVN/< @X+SdVX)"̀ܠwL)U_+1F4Cs|\>{A\2f""[*z4{f/ f%OXcGclBY0K7\{\>{\>{\ &xb(ώlT6(*/W"yF绍!F0)c)V81R(J P`fdyJc;v8at%-_n@̒rG˸VNWvss{gz]ݱ#4`-2 Jhm"IbT1xVb\%7 N|%KZ4FrKϵu^r\ow cQ{&g{M@z]j1nA;]R%8<y{jI~Z\jk"tdR*c9ZHyQL({5q?V^C3ONqԣ4oBꖳj2N@P&F\Z9rA<;j԰D'=7p$ŧ4qj*qTsUÓbl9l. HС[wwTNP׉C!0W{]*嘄7>z|Y5Y-5ӵØ%<j' VE\`BJ )’?7[WE`="qWEZgWn7WVh,aW$a[î"EJ1WJ.hrP]4ыk NgE-`#Nk /Kpl-pe\_r jUvt *Bb5M&iLHhjE=KyDc~K*88TӏO)rCmrs@d.&u@J7ӑpt;#GuDAemgY/!|Y(͉`Ѣ g(}Q4.B%nwz'ťUXI8*o}} yʇ:CܡPw; up:H?{WƑ ObGrIpX_v _%Is(ʯꙡD|4)؊fMtUuUWu=܂[pp܂oQޢ|qGoF`ÿv+^v+- hwo=yc>nRO/̐/'ٕC3KS2p cGrc 5Om cӫߪZ7ܘw.Q(F5;_ͥm3EuQ0E ]0 Pt.@(EryGڠE.@I*@(E]K Z2.ȫŕ*ȫy ՂZW jA^-ȫy7(IA^-Xy ՂZW jA^-.f.&5FTx)w>,-QG_G,Wv<ڭ{4c0}k8O[-x ki8- @hM^ 4y&/Dt8+X}WV]}>"8 IƦPy%\ !Ɩ|b'h>$($q,W^8-CNɱ$Ы%Dp@l\,2Jp|0B&ƅh8TP Ё6ٍ퍝[͝l>|-ԍ;p6}rd_^u_)׋vJEs,jQDt@@2Ə)AUJR@d7@erYP(UȅOzk,$`25Z5c+QbHؾn~ g}UtuVW)Bn)XrG\Q"˯YJôxAaNet z3;z?<T,Jq+!,QJ#+1ɓ;'t$ + '6:[t #sOQYph4OŢȂ2ZGͭ%S&"q@0Ybj'ViN}G?6pd/RZ p'H30\Un7hᒗyAq I(8M &M6)#D$ÝUH\4wYipѳaJ0f>pmrB :!B44Jo$u2bJ~Z, Cc 1R#cJfF3,I#I&Fmrwx=Yog^yJjfK bBnHNT&ќτpޠ {h-sɍ@*[P`e`G2e` {ttܪa.n:kn-s՛'8!FGؔ74> 4Adn'R*"USn%reb8z1u\\Ϟh6Yd^/a.Lͭ~fM-}{Ř\u{t>Q@)*2(9:$*(MR9ʴIxӢm= o\vŅOEz*Yx|YܩȴHE'q\ȸ&Z9k$'Q>lnojˁaqU~:9MBgM6!U3uǍ# 6+iA:>hue^㉼ %hiU1cxNu" vʓLܴzly\sޡ}#@<5-IfzI[eB=:{Niī%s3'q\x^͛`q}Y?>.J *4VGRAt%1M~bRqztr~rOA=\kǪj~IuHerSֆ5J51xtD,2rM10)׻oa\Կ׾uv{88]L[&@%-8ʌ(:2 c%mIE8EȲ|j0&@4((>F >:ŘҎ9V0ɵDw,%1 X ?G>O5z E&Zc;pn2hӪ{׽\DEFDJTez(%IA5hiT s$_ O?g:_\bzby5CDBE${s0rcCSߠK#w\<<]~bL?b8_rxwH!9W 4#J(F>INT4`y*Zߛߢ]-:{ч㐝}L=); dORF<+p7h:-,|:ruAȤx_1$A%@"uޠ޸L<7*K<J }K!ġw{yz`<;l{x~M"Oթ3õ"<ש(`8_ᔾ]Lg\t/ٲ3t<7H ͧWVt8*!=̏e?Qj/ۂLqڝ՝ Jj.+pbahf\~ '2 hxoOf$3Mʾxn\I[r'M/աٍkxs`Ew=a:?z$uuuM6Qgļ#3mԷRQַ6"TJ)VRK.<* v{gqwG싈5'ùsלwG$UshPԋx[?chQ¸սyG]X2z?u6м?x3] NP7,(\1Qb8&lVLLݐax2,ؐx{d4g&8%( =r+Mě(QFô b=7Jqz!0+DqY 0>&=zUߨoTiW`xN٘Q: B<:)14}A(B*ee :`ƣ|zzdkRy-r:0Gu>~4=3FjCm6*Cz¼$r3>=k~B(F$`uR+p)D0ڐ 1\S!L!Zu(FRVhcT21dC{#w[Y@'S69 ܂t֏ٳ퍝oٵ[,7M>yKon;&Nce\-G>?9bۦ]nCۅ'=-[{OטZ[f6;BJr?>eowRfy֟GW3=%/r {*] wv.)-z;3C I:f=; 廃#{Wfh.zh7uH9[Dn%6J90pO''7Wh}*$j0Yܻ֭gPv_'1 Z\Cڊ*AUƒ@TDUR#ucW=4!O퉶OXGF'GZI=nWWvvʭ~ %͑j}[[|{HWpg)W6UuaKu/Mڔr4go٠F.P[wЦrAZ3"a "22J*p LhSz{ x :x$GMpPr: G=ךԖޣr.!'<ь D %$䙲BpV[Zzj"05@n9L̜pX8Q$dT4k7vS_sD_-beYQ׸< NT!ȬI)Dp-Azc 7O2X6p9q+Ƈ::7_۶"8]Zѓ`ZxŠd@HFSQ#EEˬZ|'C(MO )(!ϵOVܢm&i1c1'<Έģ0`j+"_K뛉/\[!d WKhSxD !P8W")_ Dr;Q_EzhN RGh^, 0PÅ*' y9: 3=37Wl3)Ɣr{ޞ,d3% %M S ~s~o4s5uPq$giڝ,fofisDw.??؜P`?TkQEW]0R4ONCAլM1O.ZF5IuK>`<>ˉ3Bj>4V ϦM: !Gc'Fx;ukPtףqhnp=|r[gq[i. M-??׼Mݤj ÿܛ$(V=z:pD| x黆ec86^͔28EdP3j0XkDٌ]`\Qgl]+79]1Tg˟Fhϡ18j=!H@9=&nfװla7(V"NPPv1>6ҟpyfѬJ.!^,u❩7Iyƅa }LYF>eƧ>_1Nzp{;hz(|FD#Sۄ*#%T,}w&YZV>w{0?|t}Ctq$B!ccq `}9"议ˢBJ!)J֐Ք0h=5=UO]]ՎXt+ot7~PMoq+2eKӕ0G@9jR)6 #y("@XNkٽFұiZsH,h ?/c dXC/1)=BZXɒQ#eUɢ`@7::,%B:su=wn0۟aBJý-&OJT"7˳ѵ ^Ai: '4#x5[$ ^55ΘHޭϬRƆBj-FX*"z3ZI\&D.8SG]rXO8)ҫeM!uVӒ=qq=p\t#.I38>drŠpM 98N(2%37!pq(xXM;;m𯪀ݦ 1R׏~ pJ(J-v|ʝ/N? aǀA㉤78iebه y0H1֤;TMaAzcܓS(\i9P>B"+R0I1sB(1$e Dań!C \I.1̅WWg0Y \Φ? (q0$i>,ۇm;۳>4BwSCjvmS?SÕ~>f3aYK&1l2gdYZ[LG1;.c @FC&v ~\羼ëk_}7ʇ9E0y,_Y:nՙeO)~T`7{dS6Ze̺DAZz qn@.!“F|NJCr~N>bjn3|w2;)z~1;ԧOn[a*sX©<h G,D EYjtZˆA 5aU#'ED#gLQNGN͚YVGꆮp4Y⩤SN+V M)AEn#׊)rP<  1FNq/Z}.nneũ;nKDW 6Y`3שaֻd.Fcw]8ghPrj3]@jXBVjx1Eϻ | k_&Ƥ8 (Cfpa<$\Fkp1zDCndN'& 6[? c,Q'?|~I>xn]܉Sq*=a>GXꔜcD.MㆿZɿOl~εHMs2=>{͏kзx0euiMcKɥ&u|v|/d6a_;9Y:bCkψ;VKȝV} L/`w EjJSS-- TK͜pλ-Im]͡h?Qm4 Ooy}oi׻ܩwgf-{w?&ob뽷GSGo'V>أ^UkJeYҾJHt7 e2a6jEٳ;-+Mj 97 ̜\qsnީ|KwI Nh"-u,ϑ Тvrp}INM\iÍxqöGN%5\J0 Tүm4nEm#GNؘ)Q:cHDЊKf$X.)0{P~\ͨ(=Xo*mAUFg5S 2ѤD@b KQ Ǖu.`M9:W4I3ǫITR/8C93V/_P*9|U6YrCD);JO觵Y+g NsC *c΅|鹪~|MY\4,{r Q.2H\m_+)$\|͛2˵ҺZWmw4?c:nzm\csZPRN}gh;N-/2&.xw ئ_6ݞva:VW}6qNa")[6rFIqkzs ABAa:kòEHg7QQ|14%iOO^dޖ3|g߷?iJwXiI&9^]&I62"vҎ;1G;R|݂dQBu~D½>.emmnW2Txhw\'xoMIe$byZ^t9;W׳yi9׳Ph &ތؼ6N>Mg+Q/gdr'z㼪H3''GzMcd !n3*ZPEFŠH>=%65%6Ui;T^~`٣uS 䬣VJl[`L!2~̒Y[ 29$g a*OE Jq]M~CܷBv~צ7_2IC]gМ2Zk_YSn՗Kg*6|pGeR\uS3h3Zt2zFGSP7d˳uIGZ̪tqj;4x,x]QGT4h)TRժ8/?58͂#D'7 uB 9"wl0Y&^8b.;DKDὬlgZsHfg2tS`KF,r%`ήgGEהȇ>{#.y4&pwBv >2˅IɡpJh1`.)gL!jz!X÷jȓF2a9o +y eOFi[q3m2"%&b48ǠN TӉN fU=%I.vxi:qܓIX`!HCƼ $G1qMªD,Gk"B]ϰ)Eχb|Td_㓕ocr %TlPX gRX[$qz` <+ÊrVaCju֡EyX!GRm gYvio\He@Bq3`뫳6kc&ޕ$Bew.)K@cw{LcgÌaDFfjӤLJv˃YER,ꠊe$ueUEds8c 'X'`SB+#Mt;s՜q!9k0#k1Z@eÓӳZ;op:OJk3lV:U# "O5ϻVllm,5w4.4J59򀍿M w~6((z !١#tq98BS/jwELRn?o3%Y߾EԏO4QMEd3cBL00B6HdS[% +!Ҳ@SʨHRJT-hd cIcxx:~P)$'R)=R??3^}Zg=mġ b ( rqr&=e)t:dzW8(Av>ͱdLqKt9Yi9t26-cX7[؛dlk m϶ p^VDgo(6ޓi~1|vhp<~Ͼv1 _L:!%kg,H.NٗUQ_.{1*֑)9+ Q2 E}U5-vNZ7V/m= [c6BX9FEo9zEF`JXAkIXPkm̲X2#C EΎmMʂW[6I8/m8w6_qobc[znEM-:U"PĘFtV"jS}+N&pZP^-bRl@P:9Pڙ"y03=G F'CK-boù"~zUG%>:{-=E?.&eu 9-e1b 9\|%+9J hU5!?]}IǞlw`zs"*ۊYQ{R5{U47%7\l֡8ʺk$PQIv7.. _A R"P GYJՏ9:J'oXa#iqNJ]=61 *&dI׺kn{im)Wхy\q^B t6U8>{y^  q{rMxP+z/Z4idJ'B%*I=]JH`y.._LhrMQO(*+b'o@T ! (@"R{0(R+1,c%> 5XxT dM"ªVVzGnVܛȬ1), v\u[,x)O5/9k2,7qi'uZڤ&?^?m_vvqz6Z@ף^8]>}] @ɓN̻KM2TţߍpFEiU٤SsBЫP=1s Anl;|s:m^re^@'r]? dr|R5rFOm\,sYd"һS5?S/GF~V&jj/C֨Ïpupcv.!E(xtFIՇڦl:eMC x}]Dj^o^sWn|>~GBck+u}_ɽ?,-z+2qI1e2Ƒp><\\ 7]<7 QzSD-ְ^?O1uCU 3hj$q6UmVF9~UZh/E}a?2z̻uE)IM`GoBs LP6YBiDh)b5]cXG7k/Z9yG:v>i=BF(X7)Mw@^*7ܞz#HꥹZL1%4Z/l :)Y" <!с#]F<! g A_P|Z+&)BOoҫ.* գkB>>wɻdc)Qn $?ٕ@hm{nhmIWM9}CLZtjjj0~TWq%k;St{&gvagP/ WҦ( T5:E4ZذMT%BH}U zz @/ˈ+ŭ(فOCHVxTMDRV}^?@o}hAa<Zsg; gBZ {!o| [i:kI;i%nMW?7/E#~:xԀ;OS։w?WY~px*.F/G(bi鶇V즈eg2-xuW- ?ev ~Y& ]zm޷+z,-8o-Z۰vu4=ֱ- ,_r먋}= ~EyWiՄk u#|llpۿz炑; m>>=Ć 3/o nVX)O94Ć S5OpL?Y1MΑ DMJSq,ٛ4*GeAіRH2< xdHieAEp,H n8p.`8~ܯFc[Sv9l+,Zs6|pwȤb+gMqX`u%tdޛ6"%(|&Ay{QEO+z]M;6y};cT+٘퐎3cu`TMRǵN 84lj8 2gY Q̎Zén '5أpb?ZL̊5(`B`(fL)6ev;JnX-͟zF^ɲ].uy,},| 6iv10wZKs2?]thsTƣ o46 i|1T9  {3̰G83dxa^Hmf'R`:YH9i.BvCe1IC6 CL`R$-`7u_ùg.>˰\DP^k L6^.w%M=5T}TCի.ÀW |  ؃x P{hx 2P3P3C֯>Xk',/ԯ'q|Nq76QYhH–,w&Ms#V:5H ?.F+%[0B-xKvu?[ @EخV4UC.Z2\Р8JW*,=F^ L uzI, "Fla 'ً@H0|{4 SX:~J?veؕ5=|Aڙ> ٜ-KiRw+g6&`B)Z>H ({̳ـ`t;X7֮K !Fɰ@oR Sg`7v3O4]!.3_G]pScԪjuZUՖ~Slj:.>"wUѸ*⮪R:׻g*ừ*}<MKk9twU\ zw|ܕ _=W./?\9&j^܈o3=[ֻ͙HSwde݃ADHE[(*]&2T((PP"ZmJ?Z'XWHIV`l!!B[QyK('(EJ5hu).ņw8h#u3ȲmUV"贽|dpII]k۶^bNڷ[?޿JhlEg `"-+d7X>^xu5,jU1ϳr>.6ZIɛ\Jқ7c^ø(EQHТ:DI9W ё#](-mRs w3}?x8J=y[HۂǼzMrD! +OR{ FKm $E' b1(y;vR}oR[_6^^1߄hؾ-]J9ٌJH1/i[w7_].WndW= N˳彠Z#Ơj(7N1Ɇ0AB؞13GUcЭdyAx=̶@e(yORԱFZ !*ɒְFLE= Jw,(KozH 72Y[˵3^t!Xl.RR8e2)Ui x BŜ=FTioe{<eRk2XUhضI$,#Ng+BQZEP<܃pj3Ҕqn6qzn7Qeca"2baQV:BZBh^rb9g e%*#W/: " QFziDq2%D RЖ}ANASSQZCa09s+knf.3 lnxdI4K]7|Ť<,gj꿾ipN;|9vEO_+_[Xk2}_%Cuoa@i>n8퓍kj4G~oxQ)Q[Դ3uϪ۱3s#,> , >\ހ{9SќKg݃ʼAL.)O6Og?˺i2ͮ'ÿcxP?eh'tF4^Dif/r.;,~_|{-f(oUOH:jE>Kqtq*ddʋQf3" ) eM6 fqkRoY|;WK,Il,/#JlD:U/](#rU=9O \G /_yO"Tv)>LΏbEK0b4zT}:\-5g^6Ώ79ڲץRe^ ~;8μσy J{Ƕ5\-UAmXہS3nuP*ird>24c)% %@}G@ `Ӵ!EA pćՏe" T!dtȫ*!ǔA\c>Iy.T߭tׂqWh>liq%>AIzuK&Ch 3v0/ScR3+bnkߜ洃"]2:MKQ&BD6ilUK%=B.3F5 TB$/ >auHE%^HH@IUvk=y f1͵Rf~VdnPCg5~XG4Vi7V)Gc㨳UE?䡋w"ThњS*MN%E@}4늴ԣI;hKonQwtw49b2t? ٯlar4z,c<:ob-ņtQ]N28IVqypp@Q~}EGe.[WɁdNG)moo :!tDlΔpIs (UHB1K("%oT%I*036|3qnXU}@#^?u KX$퉄fۄMVB$ C$^+'2^Ys@Ng]98< yZ- [@1e+e(cFbAlaL)X+\z2WW6|:n21,&p؎|(_ɭjkᄑb:ܖho).f׺oVsgiذA~{;TsB7l3oUQ&-N%O2EɎ8+b9G.f)2ZAE٭g4Ζ:D1Z e'Grr ٜ4:ITIu;ggt ;ӌBӱ/Խ/S_^CEǛEև~|7v@Cl2}/-84CJB[Բh]I X]tzlTA 5أB TFSS(}FoJED:؝cEZ 1kwM{՘MpQ^HG$1gT 0 bfYNaR NW: P"GQe˾&e dp0p̵aXPHE'9Ζcؙ8aԏ>,&J?nm_xCJ"R)1P"*JW^AZMga9vBpdJbT"gB qb,-"+1Χ=bge~ԁ⤙!:;Ӓ[UNbum?hx ƈpQ|tl:pICf]/?L;\ز~fLGy#܃?XeI YE[x"9wvӾ7BM!^WRD!:EրTbUn`.J2^!"s!SY:2Q8"|Ⱥ3qnX9χToϮx~2>f;;B]/V?Xcgr کKE\D/}X,V`4K\D6H9EPR3qvu' sLYζՇ\?h)'7~QVp* \oŮ`%x5g6l@jDb1 f”RA ;3wqFCuY&!0b*SK%i/6: m$JunVCāg(5#<U!%)J994cڙ87(X,~>c)sB9d}&%J +IcHDR05tхdh~ AzK9 K-jWw):Lb"384 +AJȁݕ=_bB]401 օX 7։ PK"$8[W@i)xvu10S(AT)3X@ dz!#dUocdckrG~ak8!xrdP|t Jw7c:_h .O TK:#Ɠs MȎ4:]IC_Dzm ӭr{_Y (J̖c.iYI 0*JAB O($e3g7ZwH+di3xuK2VnnQ/2'-qtݸK[7ݺ8Б6 3Z4/]T^4424DjD$݃-6wp!={|=~|M?i,=z He յr?CP*}vq fRMP@,N=J77{7Q9PB)JR1/rqTLVSN!!dWI_A5ܞE9_5loL:DxT"HK%CCCuH5mFEKn6 0^.H.e7j+cKI+vlȔtgUMOPK ޑ3&N[Dy[Ĉ۷k-7y @_13A:+<Q=( "m< ᇓAN_{01?Zx$x5UzF91j.oD/Ur3; nfIz;qN?fj[5Q|?⺒^_a7x oiRN򀿮NHk|3N&.jѶ˨B3A3ëuivW9u%*ṁ`T__)͈]}o7ˡ50YZ䥢)jdP\}8 N`EBIӤM`'pg%kn1]8k/O{=;1s zPNBHYC44Jo$u2ΨҞ[Ւdf) 1R#cc' Ds,I#I&Fmrwx8;YWfj?5|B]$!jGST&*(]oЇd={Źeԫ@IxRkwϽѲs0ڂ#,U`\U (ȃuZV1W/{',2 ])oh|h4 $1Hs,Tu浕!.Sp18*.U0w*2-rr=VN;kAA/q}L#jaAq*?|v[!j(K ^VwMI6eg?B# }3eniA8|dgqJ\E4TŌ:9DZ$)OژNh_wT,]gR.SNlc_ ۘaR wju-͛MpUQ9>+ɘNܢkR;;w1PoL.a['T&0) 8km cT#VJ51x_Ӌ klMqnzSz}CzR6PL k֋k{oMˋwOuiMJt[ qaP50td@Z"1KZ33J{pr.tK2 ԉ !JQq邏.ӎ9M+gjZw,%1}"E G.5 4GGf~繻սն[l%F"|Z_4e-#QQQ!'D^9JIRDP5z+Zne"{SЛ>xz3->_NV#_qՄ_$p~qNvK9/mg7g~lI .4|i-clm,gyiA=flb},7x8SR63G%ԕAqaGIޢ@P8 Cgض:ke{~{zW{v]U3b~?úwq㨯=+{TqiQ)V麔9qKֻqNn7ܙ0>*?א4mp>;LL¸yD-[jwl6mnýbl,jX$? ̋+km}yߤ]o ۄݔ1YL̬"Kq-`quvʆ,@sDiGngQQdF .$HeRn z!0+DqY Gl_nPzֱo A8F O ƌ:ϝI8yH 2F ژ6V}::XXlmE uiW_y[uĵLjO<$ h9Vُj=+RN@X4K y8Ï;e|p<<@MocWIˏp>5QpJ#tbb4g;7˩WɿuTBK%Oh.)/>iqz#y^]rɦ.F@rCcիpyĝPp5*޸Yϧק ĒMÒU{VJZ]ۨj MIѰç\͟DWcꊽ}k p`@6VjS 2'J"90T 9 ޡPAqB褸W|$g!bBBS$Q6J&2WK%.:F=Ęй :(R,\,QY7RUK1qv.E S_sְH7v||{-fmM:rڜ ǟ[>s8ft»w=>ií'(Z5&6qCbras:v;_;οrAϷKd,z^jy=L -nZšxv#k>=d>p*[o ̿Y_4ݕ}w4pTꈣȭueǍ)nl87?aD3ZF0>2FcF1 ̙8,.gxYJ{/|m+/p[Dͽ߷&+Ml+ &UPy%\ NY};Tu" \b|8&T@NČOeʛH=:%_ -K :7^iNnU#t7Ce-.GGN]A48.g$)fHSJ(7뭔:꜈&/pYT=PqTJ6rasBrd|TkƤWܢR%OJU-&v ^guI~zCi]ћgG_}qSq5h‘qE|$H!9TG#ɝb=<kn(PPʮAe>jUsqя~([\K캩,ZAѸI?!^QYb0edSEpgJ9WN-B *8@LJՖȗsWvI|Cg0 )GtVs5XQZFtLg/,_8(NjVO#:RJ*+KpZ(ni.E0| +2x%"MT pDRQ 6 i|:;YWX겮PSKKFZcT!)6NbUe|ǘu tH=P'(T:,' |V;ǦT+iE6: bY,5"j'˨ GtM6xO4jXGBKăwN  D3 *Uš Ҍx~1.jӹ;CLmG.4J4栶(d1QxI"aY$ %a0B8uDxh"sBp! ?w BR$`*©@(@jhKv-;EFlcԍQjS`ä;3Ob~G @!A`',( İ%W"x8)IV 'j7~l)zI"0CIrA4J`aLkөd0r- A(yn$y;D=',9<1 mCj/onl#XVɞ>xbj !eo_2,B9)Cd;OAW_>Q]o6ZbvڰnJ)糓4a<[ނbddY%*jeOUh!w>^}#7!Y>_?1 6a |' 9 +)B֤Yפ@ܸ+Z֮7S?w:}"+#oxベI5 ´%;<+rL:ޣ[d=?ttq t tN:~JO U;! θQ .ZTR&dj+w %z>nhZR(| 1`!GxO+W]O>Mտ~oLsgfq}rzWJO]޾NSHK}3{s9_]}&ڰeus[2{w1!|jGO(\' -T~ZؤX]vt?l@b4"X5qX*h i`ԟ|1O^26:{-?F? ` \$Ӣ'=cq1o}T#&WKu1ٹ?RnBnQ%5Fv$89^_'zD# g(ˊ֥·h&%щۑ"vɝT۱d;p*bIVmv ,WY2}̮\g/_°lQe']~e6Y31SW2Լ}ug~oao>/Λ͢1 5j "!5c{RՔߐ;>abGky Dt髟E1Hݱ*cۤ܀QlM[`p3 `c܄P5QrqJdS;[VZd] / {N7!чs{p꼯V?|kzhH룎HTAFp'3L&Hv=W}|w# ^hO=D=s+L [G-Y=똣:#y(̤R^τuRNhG% t**NUF7pYD]9Yp[uڱ]OI^G*#25=FǤ"9̰4N%pD0AP8@2=(ۗ =PcEc`]ȟJ ӎP\ LEtH esXAz&K p丁)js4HQA7HL(09AXB}8b&l /#[FuYF+WTÒ0Ɏp΍F6@t$HgK_ ,y82yB@|9>&'ύd=ZՈjVա SL wƲ|0x""\yI" KBS7|\ _NQLlfKu>+^E(שD4"V@:DQ( ګW$9kUVd a.Imָ4hZOyy ˼jkY53OK<<["jd~l&ͳi r?)s߽ubPI$ dl{`wdpi'ף Q dkC(XK wH]a!׊\0]DYog$zqS\{zWe*!QB,l*3%`ylVs:u;H.oI.<+ )v*FTDOeEg6RpfR\4+M67 gIziT 3!ŊyvR*n\$h$*^x{lq7y@s=6Ep0o-FPBq9JaveHe s!{10 f1grH8+ !`C3]ViuHcۭ_IoSz^}(&&Vc$Vsgh՞n",RFXpw*o. Y{U]M|qِ% (!LK'8 e#0>xv]ro~^Lg\.J'WS@2-kZ? ȃgW讙/SxQruo}O+7$ҌKRdNBTQq H'ycğL}B<-Ǟyly|2 JfVIyYl{4SogəiR"O; WiN}bo&F7~M Pޗe))y-/EB׼:a* FɨOn'7=$#_|؜OAI4ŵ誨U{,b >HvǽjȔZuD9`\TDzVF}v}~|~ y;zSlJa^iGjd^ ]Ks#Gr+^vVȬD`K:,X+)q$85#͆61A"vE̯2*2٨hsp!W'E^A n}8O{,t܆vHPM[;ɹ8_ݸmg2zg Nܼe)eX+)nB㘶e!!ۂ݃65xaE9+ r M^E#tpZd0ƨPV303 j7=;Bv:Fs 6E*I"xE!g-JÄNY9̅3[vL@(U?\df#o寘<$d$ZHFy^kfBeC&nCMsƖ|% PzzM.5y||#<9*)2AfR!#IqÜ8qʚ@ r ›8\󠝖ZE.2*s{Vv1\`A"Mt+ 8R8R iơX*c!Xxt UߋggRgx/hx<<ݾvv 5`Re8 tt2&cYh́-@{. &RؔJU<ɗ)cd !eٍ~r;!foPvՀڃ`y!:(zu !XpVHAH IIn_&aF]K$$ϐEA&MX J$a )Hs>f,PPs8 {ҏCQWFD=  irZ,rJL'@!$ U 0BKRpƔ&wD%mV˨A3>`(LdBmn4PVFjFϿ%jO8[l)YMKns*h\p!Ϥd!4Wj+Kr.r FzC31p/xXM;z=aNbU&Aq>cE0~cIN#ꇵ[Ϸk\(hm_jq\o Jؘf.g+ƇG:m{mzm{m{q<]S i!;#MҀ(ȓ JG*dcx?;a#73'#`7ɪ̽ E(NY,E%sx8.݇Z7zRk+o ; ϳ/n`<lw`敆]'f_(֕J>4/aru5nS!uj L[Rs;sN2,ٲFdFsko +._mAӁCvoT .q=>ġD*Yۢ%wvt;<_:"]EoM Ͼ98t;vvSf?cokkV-l/+>zcX=V|n:Xa67cw]**VC s^ 5^(hu,2SC ckw5أkEf4.,Vg0Vh(kɼ1XbMi2Ӹe$*TL%7>&eVYX%LP\fr_M=kj57pY. S697c4;y-si]R"f tAds %3s 2UU#T!F)cjFXNއB)ęy=GH1HpDe#&n#^y] ޴W XGǧ̗J_#IwiMfy}vҦsNݏY~qː{]a01c`bēY k|ȪdC &ʘ}wxqQ{g9{9+u/?gsIE"q.\PrTYr2G) QYK ,H LK- E(I91%cR+i6PD.E+{٣;/ß}x\'UY6`Qd't:\.j+ow 2XJ,pdM@R"\ILSo w%0woP 2spCB›g)2#e+Rtr )v*FSH9듍8->6D= s6BԻ$!_%2 3ɒ'Fp,둍6t^pBpO ֊|~s2hHL:%8;*"9"S#i-}+RJ-VɁxqKObPCmQ;&XgLɸ+@0b8vA,;t*RZ-•L*W);KWo\wU>!"q?H{pU섬\:\=^3 ^'a/ W׊GI?NJ+\<fr 3Ey)4 5}dMiX'I18 ڙ]lavپOӲyr^v}3*Mw-6MVxŅ56,k,IKôߌa?~5d)&>ɴRoE}IZQ-;yO]My)%W>|YŅ+Wo+MWo_o;[e|7OHeO{k-f{,W*Ǘ$wX0ۛo:;ܸ0 bE`kN#qӧi+RJ=bo x.yqo)@B"kmd3ij'gygiD͖glt؄mkSHswM@G[3iq\`N`&q fһDi7^EqjP״49Aڙ}T-uӟ4WBj`ov<O_OhM%/mp1ߍ&@h. F2d M;^c;:lqȬ֧dC3]D֓myxxȝѰYuݷ;T-EPm΅>O7nZ¨R5?~t ݑI »Ro?4(*<9Y]`yR҂(60h`W8)|)#x79C_2xIm7kbws^BnPdV*UMwx)w7{mzT_k+e]Q k)?5^vFP}*kftgUx!6C+QB6|PM Z˒z#,> }nQg񡠀!2+ ;xWy\qy+Zq(IGUFwHTX;8nL,Fwۂ\V*(`wL)޷v~iL./7p>b;k.4ڻ_r&Gԇ #wCYRD&׍}PO8WGR- eJ"F5ߎf.on[ӋF+m Cw"\n?G),&)M'os ._mAӁÓPpsBG̊NYWS9bV=vqӏ 3))"+S"=\) p dwX"ٞGq&vSgLևdZ&M ggQ-!MCRBm#rO }nRdC92Dps6T<6zo{2?g>7ݶ}R5#{ʴzZAq:ej-H98joUCiЊ+X(UVپU8ہ+xWEܗf}THɇ7 WR0HW$<HtኤWkU٥ κ$qEK5 zoG/6^fxLkD Bq|;z݀FGO]?keM)W't`#Z %dR),5t¿19Ic!yshz\53#TRmm3H!ؒ5Ew:wV>%tls'Ω9equ$~aSƀjZAi]O& VssX:gqdD`d#=DBi"iK{͋jĦŜlE3Ä>vyPlR͙%E&@ڳBAnQ:K+v o u&G˂B i%6y`}X[ =Eg; 0x X72jl;I`6p(:\[dV d>JɢR}76AJZ2I$H]F2Yk)A@&گ. JK[PlcfH57]Ґ?<;1m1 Dah $(  K`3puk). Ab΂G^' s$ ;TkPI!0<'n!fT fǥ~3@f"6@@0 :jonLEwo %RI9n|Vkϲ oA"=dH_(u 87Pc4i<ޫ( E<}ѣ f^驈]gHH/!EjuQ"zJ(r! eA ijni؉K=FPQ>6ic hABrGӹ~Cw 4#sl 94߬3(hvN֢' 0b~;/lч1!yVBӨ L0b6# X 3 bPT8xiLd/PM6W%s"!jmL fU)u1 %p&.,J肸QV24ITiy],CSq;ff:Y偹Q &(1Gt_q5LnΎG⭐!@@8ƎU]ܫPt5.U;o:@vTVF]+#H b!җ:0Kv=OOd~3d&Ji:'Xqo.BfȈ2!A]ژ6X{sP6D2J@݅Z pu $$lLCQ^/UbΨhBPL uPt  R(U_c+PK`y$~g$+@Ak(Q0i( 2Бk"8^IbePZB ɀ!jPk M@E7gK{a);L6( v8+lA A1+'4b@ `\!kϢ;4DTL΂P< [aLAi,Fk! !P}AoDvf nf!z3ϡ, +׸ @׳m>GO^TIg`>QlFP6M @(v6xY:Z;5W7$֍Qg5rFCo b=@98{2f\AIK[ҐQUrC0C{ _s":^'%\cSE;8F l RR=HρOlr <XoYo~q >XDmi2 zX,;y.X j Q 'eޡpnE5$qQGbD*01r֋b>-йgQWcD54'E1@T1se6@zX{zPҭi@d]LzĤԎ /Z)'5r]J %/@Ĉσ(vy Z7~|\b񅤸`9)4j>:;ˏ''6MzsxOo2/~{^J??;xֺ{ o5sn+:Dwm=>0~n L衋u7=`OpU ȯ "b ɤ8i;QRP'+tEw N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@j@P.3&WpIh}zN2:uF'C N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@zN 598L MƮ JbuF'rqR': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N u^'|tkr!L 'zzN LN &Q': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N u8~C6V~{s}z &Qɸc\ܘVc\Z/޸-b5.ү 8]]p2p&t(9(]B]}ՄZ h/&JRz*s ϯ9yfz\~X< m|l2b0=cFa#VnrS xsA2sb`[Keŋ]/:o yNa.(z#Lΐ|NVDƭ'\B t(E_!M;2oюn7g|2oK#uq[hc_c*77wZa={{Zw[\p&~?1ȳ~۫M0^?w4/i<~٧:9~U]ޑ tN<1ǿfw2̖ʖ6YyF%cd1#Gُ+ǖmEJ ĔJX(`G~5pCEDDKePQE!_`Zjo;?&J"WHW,&GK ׭F]ML+]B̊j)'\Y`Dykk롫 &g&0QBWNW@V5U^]M~=j q-t5Ѧb&J `|`Y]Yhū2C/o!>g^zZy&ԃP&–t%JWzkJW0jj +uּt()]FϪ/3,c9`yB$(q3uvcVT''h^:MO6*MvN$ѳ{оڷ~92O&yӇ#uS~hڶeP6 Ђ֓i~_R]iEok)2y[Srw>iI|Yx3v*ےkܲOlRM:dٙ5?gHk@!BcP c?of)b9ٻ6$ 3~?ws0v~XSbL JUpȑħ48l6R4Suw=atNx%(,Яn1OOo<\\JťJN3هg-?@pXv"w><-W$_ك:K͑ Џs^v @E~67nUY'-!%ߧJ2k@V ePǟK=\Q-5J\!l4< 0̓'*V2/[м0 l]|_t%'sh_Tl.bgaI¼blm톩eiKC{*07E51,y\Ƹj .o甄VPv)j Λ ACW M+@+kq(Y)ҕ Tq Jus+Qc&:]%Ր]I"XX5'rBW -fNW e[$JvĘ%\4ZyUBZ)ҕX4jBvp%MRQ ]tXE!FLz\:.m:#% :%u(]SI?IБ;p5o ]ZFUBpKWOn >}Cxe>B]0%"b۾#vzTX:2U׷**qNpzVT>A=ۜé"&-n3 i3Rg8ԝR~070A1tJh/5&t1i*1tjR{ӡ+$M 9a6CnBOJ~yt%8H5`.CW #':NWtt ~h Sќ`+wpKWwz ޿ +zjҦUBо*HW +AO,=3 O a=S;m3Ҹ!SW*TjV xlv7Mc:%m'Tǰ>:x `ħq(XѭB{hS'l+H?bacw Ja4Tqkv:!9'b8|Y޺pap7ߛ~<]ս[\7O:RgEj$FDQ\'e] Me}|\Pޛĸ3zN G(2Jθ%h0DQAL0lrpT\%'tMSpF,t|ݤsɺF..HdJ& ~ AY&D4 YfLǘQu0>2_\Tt#qT`yiV0TjMzߔ?`@^_X>|/|)' U"]xk+97O_VUVt:04~ɕGirWiKyEqF~їfl>T6 JJǥ]+ {AF0N*~M8JP"'02nm-6-]oς[;Q/! 0V$ڃ{h uRNM֡cku巢K>[S-aap(`Qйzc912ݺZZn>lIw!*˹AZD`~-FP~Ki?t<}UNͶ5D>0˨ GtM6xO4L԰:0 ?fFGJ)QG"`"RSFDD bQtؑ2&"][oESk\g2qJ߀\p`7;e-@14+w.&fbxX)1?81>`Y),&1 Rpd47L:˽$X@ea˧%D 2u  (+a&USA-Q:TĀ%T8 yәLIY,- JaظjFۖ63<%@% 8H+2> ˜a#L0@Zr-ܑyn0Y5(Ux!=by2Q* eKHja3Ɉ@Zj fub8$ghgv LȚ:H<kzUJ\og< `H!RHSb){i.{`%N, &kDr8@YpCpoy>pc *O,*aTSʼn  MC71[_.q V*{f/8_dTnZn;g`Ӌ0`O9: ƿxz7^C`~]7XZFHٸ2`&zn{&a7L%5*u 7WRȵJmݡfB}(y=`4S'N&MPE- |Cms(X,g,暆nyACKI:Ac.=6[%ʃ$lm#ZQ8\>XQ*ǯt[HMห9Xoi56?#>>ΧAr;> F/}56=9}G 1mV832ӫ?_^f?MOu|. WWΖ}S̯RO9%b2La#'?jbf_>VФKo>ߧ~v%.˻x:+hd^< /fz}N0F@{P$OЊ&.T25z3ׯ/8ʥ.fPWyU}rXWPq[TU{'Wp^lyxBbab8g4VYRAһ/{a+ȃQ(Yl÷SNegw~ 8KV[]0`_@lE#]P>"Ϭ+?~Y{&5_~9Gȏ~w׋7F(ʔ-b&>桶;s(х_ޒc(Y4{`L'hXR@\/f{d^/p5-`2\V֒CrAʻQf@R{Qz0(3ޥD_BM,VWjNj sGN+핵v˖+g0vdzޡZXv@Zυ8<ӭ(mIlvJsǢr>KOxᏏd$7x(9\(JX4V<Ч@gqm8k_-xuW>ha }P-~Y:&۾o?u7y}[^VqvjhFKoڅmh%ג Vjuq_|6wN0K͡+er{#FFqt9=_+fOgWqz%F;qݟyf~@ v`Mހ 3Hblj_*MN^ 9%-ΗNӆ@*N% hi5̀BRJ][L( Y>!JvFz__G5kzlYTd_f2lX 3  /PQ[c޳ YrA'21;/"hs4)j:qIՆsgUj/ԕ}|ze󂌷()uXG;xw;48O/ŷ;м0 .hx$QdsL" rථǖVfPFB66AevKRfp~:7^ukkNƐPY8qT+\UH) 3JI7R/+m0x8C&dȁXtɐ-EHǬ" DA ~ؒkqcWh*{D3x#eY`s}$j!@`^ҜK7ąW|,P#:ӆN $wY.K3\Ɇ@LZh8RT#VCG/WL}uV"gL6Ns?`:;qpD4]sQ׆Ն,ŵ7MFگ$_FrZR[^ڮ6?LCSܡW\6= Hf[˵#9-:C>jFfMm14ʩ`9+Yf>dHA`x}/ii)o:}ЬZHXi.5R N%sc`usC3"IKUf{obp_;["}&qiQuc>%h$<$p Xtp sHC[2* -9a02ei ?X,MHȤ6"s&Y  ^M(p-5w\\ٖ$\faHu'.E4;g<,t(q+A.?J'󁗐*%ZBCoM78`y'C?|Nc '>~7q' KG4y&9a갳˳Ŵ+F\tt1?E &}93N5NPڳ$Ne[8=낀%L6n9aBWOgR; i*ȕ{7jnqځXџcFOJ^ާhzf|21aH~Dy1gq i4?[C淄V'8clGG]{RF\+4_7^ҍ<?/M%dS4bŞϰ](=fM=:ˈywZR染޸0䪐_w˰*T(Ǽ3/>nh1ц颗UxF[def9ACF"{%&`S͎5n17?Lҳ(ֲaG{eʴF:sF(s{Ƕr⦸Ms'hn_^ҿsVCQц+A3jRR9WVct i=5Z=ĂVUHe&c5=#h@o0EkO!&GR*&~4W 4|5)NTkI\[ n\tI.8 ϒKߛj?Aou/r ]*ơ SOAU%,Enӝ v9ѣR }eeJMy@+ J%,Lh@)A EKueoyxCSr^>w|S7Pw7[[V|?e",^Wp$6p.:z&tDhC,YJiY5@R!/,hW㌂&X+>9m"!~*\m`6po_ E[R=NtPo4WR)+˴m lh΄ߒSqRRMMM1|3G|v,e 啫IFaNVy!o5:ϫƛ,uL,XZ/1()B\AA1jù?D>q6Sca?hCz`gi҂srt|Sz:тX2- f/,hDK7A UDb1D:4oCM\)A :-$Ub =mJ L0%Ӣו@Jq! ȲRtΐ5 ֆ,7Gm̵1b܏C7<7K8㩑/>vy/X̺_b>t81[[De6B;|,r Jb̨f[s=Lsi{3]p$9Dx`VjGHm&Ì Fhexd,&+Qmiń}*mr:s/|}`V#vJ̵Tk _tuQrHsicUqNʺP8iP.ZڻusZW fυW֬:^~4ܫKlԘI/U!Jb#bㅆ?tNL@v?@7nީ^ْ2.ZH¨F5,1el z]YU5QAʒ[f3N t'nDUՆ@Мͦ^r{HJ;buw^J;1exP%h! bVI6!jJ2=c,0L@l 2cdƬQRe꾁ɼޝr42=knzae] ]Uh8-~.7kk Y,O }*]FPͫFͫ8<ӭ(rwI,NiڟH|~>K||h7޼eǍRxKZмwzFV2{h.y֮׆oՒWG%|z胶>(D'l#mm.wU?u7y}[^ iU -oXz:>j-u?Znmj:AGl2`h|'Zho [`Gz.l3xJ`?W(p畚igA1n D3wvhT~ߧO5CIu iƹJĠl@9K JXv) H#FqhDٻ6nln|{vsmm6J~GOlclD"Nq>|琤$&zlNG4H6FkktP ÁУH؉]$Tz/0{e\PP, R `-<"w]Pt O{]c !| (a<#Z"q:*9:9J YS[pzBDHP>o1CyT'fZ%ka1XΈ~I{US :uC!)>IS',i*I;-u2I4cX|l}~)c r/5jkɠő^ R>xe%C7w|Q~zWg"E"RHxb ')y`NE}P/r-Gz; $Rp]ؐeͧ0UD!挷FfEf. 3ۼN(FPҦ XU+U[*ռtQ.=E-+_p]trh ]e4%N]"]IB8UoF뢫 W BNWeeaGGWCW X* ״F]!ZEJ+Md0jpUk*5jG@4S+CmRWؼ"*e-tтh:]eBwt߲9To`uEG^cP"c5t;ڷ)H*5t{{tBPY6 ѕ梲9aL>}ύ&k7J"FZtkd[h*N%M M3Nϼf&Zpg.7=V!5}q<'!$HP&ôC^?V챂f[-1\, g*@hSKpHwRRɓYDm»NJP\Xdp ^kfE4 ؟8iJK֞j"3T'( bX SnZCWmV5>V(+YG P/~|>~`~~1xr^ئ%X'9ʃ:`~Y﫟F_oю)JsiLD Ž^XLR>_syD.czte gwq{m'YAE.~l D4X^@VT wˏJ+:}nGd*Cjo-d:}z2XV\ f3 +*-ZӲ@1wཫW577+*,/sq򀨊}k\.O[8Ͷ y>hTCH FexQ6m׍NT0!ZDWiOdp9iMJFܳUF;:ERhӼ}lHkJqN}nh9m:]e)ҕEtkpۓch62.$b"BAk*Yh6*]}9t[v= Ͼt`^ƾ\ye컡5GRW;M[ ;ttoU2`mZCWyY(UF{wC);:I6,=ڣOYE\o\}3rqڭbهSY`ee*6NC/sB*uév #PЀKw,8kpihn(u'OQ sA$- XUx[*mT#TwtutєUyk*5-thX'ttut%ED00x| gWXgy:{oGEVkȆS1 U/]o>q552gi"I)I0h^H\x:E꼶 ]RϪ!ie#}gUg۳j䋷g˶҇N/sߥ:g1NG 4ŃVae?_6YW* lB0(t>ƴ@b ؾp7=(lѪƿM(S╒UuWvVEV-Vcu 7._J=}VmٽG77K l&~u{[]l~L,qeh+?eŪU$?Ņ%FOc4Typ ڃq>9C$Ap8a( r(K++UO9⎟&8Bi,BI?eB";2ǃLa 1q]6 &F2 ,Z#ՁHOTj{d8.r9l+ʇ.,Ƽ˾ lŮ-Ӌ{Ln>ZͽPN>g=hd:^?P[D` RH^q0$ƥ1^VR(g 9:Ճ"9P*FVG#\ʲ~{,r6"1R OF%\U;~vԆل9#b0.j0aLٔJ)_D 1shcM*8 "ߠ3mPd8oQr41 !ov|颳\9T^/( 䅤)*D\v8EkIDÑK4*PWSRp@iF(jɨQ$bm~fI\1Τ!1R 1K Q! p=KBIǨ.ox~V-ߚfr?7+ u1C/K۸0V9k +*9ZMGCk$`4qvŦPd$At <*&B=c@ 9I5QY>ddMď0CNh'8]2Ϥx^]&lˁp f'jЊ &#FqGJ28@kQ&3b%8#n&-MqLd_5;_86 V3;ĘD 6G-ѪD(YqT4PG4݄ac_jwf!zp7G> .}XPf%~'zi&ozM޸Jb@Z|Q9/í1~<,k|SC+_6Mm :6R6r+ Z3Q{M(UDཎ̀ն[SA Kzaz=>w~P|]䋖s}Kkom0Wn_nbʂS*Ph(`,*X>݂NٻFn$U!m0%bN "qF ݯح%L60#.⯊*Fn$A"D"d^>6A0|Iz ##@bLiYbܸZlNȥ{pCn<{J 2_,{G" KW?sl!θ+Dž9PD[PIB:Q(  aKἇId]kp`hQOE5K@, DK%xBi&!>w{3 bb|b0OwtggTMqq-ehӜd:;3B%¾E7_/Go$+*H=?kX v꺸c铛`_K=0io9^Dq?_)nԢ~c KHv^}{s9%yvzm sS݌fg '.kf"D棰ȝ4۩ջ'MO]""mE׮\\oWlM!_|S&.Tލ:yz\~o?|cqFzqiގ]+{3D1\tDD&(B 4[w9[Qa.bǗhԃTMGdZAafHPs3*?D77<벋Rcq2kigL/)\ދdA[*2a;1 M&iȪmxQo ɼSLd(D }-?WNDS]f ogze4Ll"%g8&9 @eFBa_!92#yv+:wFtfB'U?уM 2F ژ`]ʃV9|́. {V7-Te .>+Ϧ@(ՆZ9mja ][$!ė Rw#J4Ǐ"uNZ$Ü 5$m?NREQn$e9KK(:F! X1>"I2TՉKY`*#gǃНlvgӻ:erEww];]Gmesnـ6T:=_fT=_wzr=k\ont|@F:L]}޶ݴ )\Jܘmtmo~%oԽ|_sF_Ro}#{;<݂#-ϓG;;e1/Xo~_t׀̗7ڄ0$O;J_S -SJ7Ǧ2hJmȜ5t! ~ hbѥ7D:=NlYTěz_}g8@k<{.0/Àm|)ء]7Skϡ8NEpgJ,Lerít^]1HAoP$*Pʲ *˨0HfRLFepKu [JH!m}>LQhH[{mdWz4dᡇR_E3>\2|P^IhcJ0, B9P%Ft u5׏.x݌B+}r<|Qu>ˋ- !'U4vr+هuCߍͿƓL4Ə~M4w)֐+. ^2kw%мۖ+^Q#9 b]~aFY R(&Rs}K "\ 9@gViԫx^9n @ u4q` A$@ [v+kGѡvŬb; BZ\ryh}UލZDZl7vS*sZh#"ʒBYv(tGOEKyW~&UAML^48h45jýIJqWҮkDySDzQjgy/9"qn29!Ax*`Z a4ʘ(sFF@e왛(pg*O΀d!d8~4 1IFyrD'j1rOEfG9@d@Ha?NT8&.Ry{,4$Rsa1rVfx4:})W"Qq8^9Rb ;X9A<9c\2U2/- IJͭ(pDE)8Iiʃ,e9 fD£@M ͝Q9%CQ.6}Dcu="% 7q|Bǹ 4T%V[uDoCuR0 8 G-: O0\<\<yX;D{<<+UlV69 ޏ/ ,ŽQώ_̀ "lmd.UvAL0h\qD%pOH%r ,;ft?ZSHhԉy/KBgI0LQe3KäET) \h83q;\FdYҰ9;z).x$.^gJ.%{iٴoWlo$s^5òREgQ.f &2gR цdVqn,O/z'Գ[V>YHD"&HNcDbF$x49d!ǹLb@e, {^ SrZڷ!o|pYoN6P)֧\!qlw8(s/Õzs뻔J(!z6I"JQ_'Y-8 R` CdprIG.Z7@ l[ l 4,܎^Yz5K^RT+ro?P*&)8ej=gV0ߢaRρRSٛ7OfobE h$ :>YCQK1*Y嬧EMT kn#WE,m4hJU*7C^>׺JzS9 %EKF9(r(1@M*{q3:F:6ar-b7 &pŘN='^X,{XȖ,\|Dr^*Ŗg z?eJ.[-E(o*ŖT->y!D⼷Y#mrvKJK9kbtѷs|,Cz_]Hߙ}xث!ʦ \ [YĽvZTkL5w_lw45ijDe8ʦITm>TưvEDŦ٪TLxQ[0hJ%6c16fb-YИqfʔ(xs)9N2mh85"zC jfYV^{׹}c `^Io,1$1ut;\2nrr}si63d?=sݙG&.X95隔Jh֙/3u #ҙa:"[z-z#J6Sk1$ךl,LER%;lu({UJl| (icK pd[RV^"G VZĴ=WW<&u.9rhW/7^!ry:ߖזg?g:Tߗ_ ϪNzuOza3^x82uw?64ׅa~^~9ƅ.Iw!I /@Q%*YģK*mGTڎB=0! qQR\l&geUuPP YN=ō5&6C>).F:+Dw%0 #\W|%'=ۻL;L::Ӯεe]uٸkbe$)B$VF4N1Y|Tp )}(of-3R=!12D}\2Ǯ{]-!Lih*$U^-Rn"r)bJcťbAynD7(\,~Z0@`A*"@VlUK6Dr0,e*T\D9Jͪr$9"1 jVE!=XU`pVw$i)}9~k$g̷Cٻ+0u1 J*$=t-{P k}+AfO\] Ћ(L&ِaZyq&ot=p{יUb:}17!ԒjlZićnZ0RQcb(f*б}lD>vUs_\>|N}bм@H e`PL%wG+vٽ;JOrP+rȈ;]ij7:엮vC~?tJ]tԡ'jBtNUK<h;]u"3]"]yKk+ nٌ}֔;V\[z5ŸϿ|tuq7O18Cw±u&5}~󓛭",*?ܜ /:tzq6m]|8/o`ueV8;p __"|/܂פx|~]~?nj4pYtk~>~rynwD2g)sZ݂#.aPrYi90A~DgpQ/#'yXtB&wƎ]vΚ5,O:`&CWnS+ܼ#RLWHW+=!ꀝ ]u^O߼=]udf:@>l3pdntANW6(LWHWS_ LG]u<uUG1tu8tjJn2t6 DW-骣:DaJkW=Uh*t#W=pw8t%zyT߮w 0} .nh_zi`]Rh@W2SU+v'CWw^EWUGaC+o]5pG o4+Ov*x t Oz245h;Jgg>@G66o"ޝg=Xby:ea[y/(:2_&?JyMck9X#\-8H6EV:5a#MXHM:\#SV骣3]"]Y ̄ NHGFQn앚pJ44M:\KS֏g"]^r'DWt֮:\6Sv_vCg:DxJ U˓ ;Z6rlfz F&DWpv1eBW(LW?]G{ k}˾LwvCtgp'FlwہLWOz2'EWUT hRH#>3e;VrlܛDz||#mS :F=y?ty4ȚhQفuXȄo_(WToGq<+TvzBtƘUﭻuntQ:c&DWp}߭ UGz+۫ & vfwk;Jo"]% 7~2tኙ ]u2 "]9k+ t h7m:ʍ3]]yQOg Y}QH:J3]"]a)wlԄ vVM:(;?]GD0 `Gjv{VW{:iJFw+?SD1 ۽ .UGbRF^~jAwޮ*|,+'`ձѴIx^xO믿Et`s3\io迗WuZؿxP3Wmζk]9tZ?7"_~X+u-иx/tb[]_}lh?Y[~VӖO׏dot0կX70߸Wֽ[[_|kG)k>Z(~)ލ+o^0UذeIeg##PQ_H>~|9{?[8 47Ϟǜ);Oįg34t#ϡn_?ƟϺ]AR?ȗF-?rB}9oě-HR`1*UJF(m#Ih/y']Y-& ZWho'ט\.tGhjqr^wp9nRշLT ]7IdVLoEo ד>ߗN11jT b6FŢ0i)Bܜ*S?cG~[g짙9nFr)jS\ 9cبcS\"˵d92:DoEҞeHX+,T{哧Pk(T2#s%fV(&qIVZoB0!狳Tl5ksIڈkV)E7[IPR:ǪT > 20X#l!E,ش5Cc6%oLʺ8sVՇSJ\qҟ{,`Xt li26!i(rh6ls=S5P:jJa0 &?)W 4ʚXZ\d'hج0Zںyx( B{zA~"iū4ew>fe! V!E+>%Yw5Ĺ.Vx WerƢḴ6g-TT#j&xD@r-PZQ>C߷s л1H DHqL5&o"e -v VI` AZa!!s`UJ68Ғ(A DBfy1gU\чUSRA/b3K 3V$` :`О%EMc ј=VB\]Jy;a*VaLxeH`%0rHחdmECE EG|(h-0TS+o5eܠfD9rBV$)+a$ty2C8ӆo5e ,q)Yeе o &XL%eU*xTQ30ν\V:LJڬ`6,zWP}֕ٻZ8Z*c:m`{A!6x XLjX*1+VK8l0`*& D?{ƍe K]gPk+Te6eW -hRER5=T$amѺ(lvŹqri_yMMNRO}Mcbkigha8HlDX/U` Do(ownXsuB֊X֗|]70;Akg-f4١[]bY1kMs҈(cdBq"Qh$%!pB 6}ofU ۹ucBFyV4Yų!m6\p #Yݦ.K6`-$t>D! / uhL`.:>jZʔUl☌pv5| :$_WHE]%74NIH2ӲU 7ʔ!Ov5}/|OA#qѡznuMq[x GWp&\,©^_rhwXCAטF8kԺJ6"Н)M M 4Ǯn6 +I HM6u !?Gx2 "ݑȈLmUFW7E|alEH,v $c]:B&YDkis `%͐ڲhB1MYJHl-RPٴҫ, x "nwEh _nAH@hd_ی_^K&C6lɵuC?躾9t ,<{M;Ne͹5e&j `ԭClkiQTL1v=(Rcf먵-h֘k>Hզ(>@fkA&f ܍F>"w(\u,jH-lb6"WFEDAL ;lrg,dBS{p2q MѓFnEk*E!nzo^m6$ jZM@1`>5 ޱ4MZդ4wSR,ݷS@1Z ^\y[(ˏ0 Azg 7 0j[.?f-n3Ka(zj\H)NUK?::*7|(y@DoY"ok7Z"?!b`1`H;шVt+dr:XTp:}!Tf7*x|2lv|t$(+#*hY|L4Ͷ{[b{+nfwBݖ_m QǮ[xHz&h!VWʼXc# J}Ry9v M ~o@?=(bJ#\j<#l%U%{(ieb󔤻+h:X (V@Pb+X (V@Pb+X (V@Pb+X (V@Pb+X (V@Pb+X (V@PjPJjWg!?Gf_PV"V*z#<GIt>qn&;<|P p " q=[lʥɨ-?WEU6ƴ !W:֯>bNT<|4;ZW .}Ftz-ESA kbCptxvOKuʀri_(DflDFYi$*j]ǭSԹ?8"=KS.BbT&BV>"TƸY] [ޒd%-YozK[ޒd%-YozK[ޒd%-YozK[ޒd%-YozK[ޒd%-YozK[ޒd%-R}[pO-JXoUY@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X x@H'%b@5 +jD@Q 䍔@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X Jǣԫ/cZjJחW@S)yXw?H#D 0oIkHVp h]Kaac.5'`{$Vt"^1]=B{DWuo/tEh݃=#]ag ꏮ\/¾U7_];v}8!8]_aZ+, t]/P#"@_tEpk/tEh+B c+֫뭥/d֯xv{iaǃVv+x5֫tf{lЃg֌sJ 5ǒXkK;\o%i=i4: n܏iEiҴ|7|>u,n87 MH2!8  d Cŭ6"@$d6Z:)鴚~TML5ڃ2OS7#*o|U9Ǖ֧߮n"<+ˁ)fD{fó3|ZAVF,M+ˏ?<;YjUQ"WYS%U%޲6;:.pӝƘjbrerL>(\YGKg>ت n%#^_꜉:&>!_?EZA[MGe}[d<&ߴi4޽,7q\1~xqv,Kp;^O?Y]|1{s6黿ђq`ۖe~A>W^^߷/W,op#97W[h[]MuR*66|>Ώ1z 7.cu7 m8_z>Пru+9E/L}rN/]l tٟnXܟ+ikmbZzfo1Zez5|8цgeڽ~bb޳\x>.Y;AvCAΗQQWw !5X%Ӂ4b匮k&[9f 햳>q  FMqZm֞N[~.F֩J3Vߢ7Mmȴg{>U]D>mʹ-kC^+濤0p~)kb:&m;e ^ϮP^])Ue䓝kqvBkQM gX.Гl vO"ҡylyzH{sx'DտafO>1e꺽DTvKm-e^tegmJ/oĐxX5qhy^Ka'm.K$0[uEPj>7j*`]~i& D3#<3?˖[yNDSG7;M֙Yl5 e_ "*S*Z#;e{RYAPRiC2gpz=G׷⨧n=,xijPLoԡTZܵmYóYMl"C8"3$?{Fܿ `GGK" !"wXTwW:Ӥ쵃|TzX))6X93=5=UuU5D3N4JQ!2KplLz Fs(ʁwShcCxV\l*o]K1xVq78Q wm4Gy"Oz_VsMkV_xn<GJknv%oUih4ӽSV3K,:A@y\0 jm<]fl_5λҾ 1^!NtȄ׹b[UH^:"$(&D91'10I$c\d%bLN1mFΖs|H^Pp4|_Cg3N>vi|q2\NomM-ўG"a Yfv +Y?"Rn!Ah++4LKbezv7?Jvkg5s{s؇Onڹƙd<~`+C=87xT63-\ aE@{o+MQlYLrp3 >i )`ٽ%k~Yʚ "$ozIZIض/?)l,DCM:qBPAyrMIA≜6[Y3r^>X,yQkwgmDtQ$^%&'M] 3T!3 ]z;dsk>ɫin3VuòS?οg#? /F,_?̮&\ ?E 'V{>I.<07>e]z9g<,Pz oIǓӋYMv8d5N#A\v!)u>L-G"},u؍亝vѓrd{ߞ=dc`gBY+,*g2,RZKm1Rim0途E1&B( mb IFX{_DT ]ZwSkFf v\H>wn_0IJlVk{~#x^Ss1sþN;3t>UҁԡpS°9++lF&TS-큡"GBauDe^R&gkhድZM<6#[*JC) oRϧ4QM%-Dh!fP^"zC՘Tl5FiYJHd)$Ad!qf𩱠6#gc~ӹ7v#>ٵ7>"k֊_V ZysjOhK@F!죢"#0%MJ0(-xDt,56WtMa1aH`&Ǻ&HVu:"SM['7#g>nU1Oߊ?vՈFtGxԈqEJQ DRRHF4)aL`FI'% #RS:# huRgT=sLLbɧĞI`2Ac،WY/N{2RhQ/C_YSIN8iKp\G"R&R-&"e"{IJUL?}Cч͸@bG=W}_.%_(5ͳQň&ۨQRIySbE)P%7o)RűQ%h5K>}0Q Cre4A>iն{Is+%=!㲕Y2S6#g,ŕ6=V`m:>?]ZnWlt'$cЗ)3$!x.lЉ(@Ml7xjZe+\. W+K@(>I;)KjS%dAcg(fsѼ'&(uF#餑kB_TΨIhru):(! tdbQ:rsmߵg/bOB Bj+xYK]-:Q Nm]!˲HP೐^[A!UdWXJ5 ʘS`*PNXR-%Gһ\ _2aѴ^j:lRYN>^%?FчG(Jo;~qf:_%CuPL%wn3'Bjqΐ\W}DiA7 Y֏׻#j5byPG9zp5{9`oB&ĺ{Oem֓^b>i ~Bg}/=DkC78 fr3vPQ>>*L>u&C x}Ej\oZG٫A7ߑP'SzO36rNhű_o!(1 At?^Y"7䉼H,X闪%h4B׫1f"칿iK EF_&Ǟ=ijrTx0cN4x?S aߋe2_f|.x5{Sg7wh/a~bѽHOĩKUّ{17=]Tnfe~g?O7qμ"غś{Amp{n~؛Hg:6WʢC3jZVWr4b(9#p{nh lpB8DDfk%e3iKR&cS4DՇ)[g|+Jzny\pp%x\N5d0tE_jJyЛНA)f# $BM ]Q&;+=uSwS+2|ah v8}-c^C.=\r{G xd1Rh^ج3I: *R.y-X{lʐ݇3V wGq_ s# jo4FWM 5TNW׀X}|]S_FoG5~_`a-݃#g@Ro$?#~3w] ۤ['!F|Ojhj xwx4f\ 8g 2C- ԳB{zG<izb>D fo99:EtZر]\:%BpGxUیx Kz$#^$ f_[ HJ1lg os0QL Ga-JC]7 Mu="BZ {CZ^v'~ vPwQ+' &S~=ZsѢmH( uGGINAS]Vv bT)iEeӵGhrGEdc^sZF+kPa n4@9iU${4#g#aͲsc3tjQ,Vyg'{z<_c\}Zf%y_KӤ7dI4O);/yt]ήtvF5ʫi nY\ n]6UxhPdDJY$,F7e1x1zk򗗑-FSrw!]dm ދtin7[m*Zq QnDEJ%{#$U8G>:N2C]r ?\<%GJ%QÉvSñQñQñvPéRn'2\)AȴJH2cbyzS)Q6J}䡫 ArcwZ{@_,{ˣ.Gj=}D|5C㖾4Wmf8D% WiLFKޚgδ-Ǚu8ӶgVLQsJB!ӵ/p9K?P{T5CMsZ ǫ«y4W rrKQtS`!JJD/ u *rErbod9rh]llQll-%{)M!KA"F# ZR8cjGg(5l#!<ZxQL(rcJ0Wh6PD.EF4J]ӛ*|k)݊8>'UY6`Qd't:\.]*s'{EE9y3k!ďîtn~To6!Vvv9ff͇_&wwkļDoRDM5c6gdșs:E=$SٞCrxgU+kwI3bpB`PґQ\MoY` p2"{#W;b:%H\'K(P"t&VJ"HTaEJmD&%J쌻*銻*jvwU4л$o%I%0Co3>S\uXgJkC<)9kɖdzs0Z+pwUwUvwUD1+cI+n`3Pjqli-WW#rqd P~`2/kir?+.iXgܞ81~dÞհ(}0u_WFr5oGdGix'3󗪹T=@O4_h~MI?w,Ө~3ݥMEtσ竛ߢfi8?/hP ˭~' :^%l* >:WͲ]L,bJ@*)YUYsUЯpFȭäL΀"."mEʵ  j$N;j$i ":BwE@]됻"%θ"쌻*bwFk^ﮎ]+ذq**vwUл#tWJ)-*+0wUH){tuJk&K >Lq⮊F]xjX+Sh* qMgrW$-EʵwW㮬تK+t]qegrWEZmwWEJW~wz~´%h}pw,|IwVu!f*@' gu zA%@|۴$FӐ~L0WӫQOF}l^R4l<],t_u & )֟nӷ:4 X+W/osGq=}Ïƥr5+7+ι=ف2:e?QyۿjIvT0J":r$݃ rQq?ҿ<> ﰐd) NɢN⧫OMOwfM_p6Wn8y}:8JazRx$~aj̶C28" bZ'PW65~4GLܰ k%MLg]-ŒYBc %ʣL]\#kq7! Ux4Jj21vh-Pzߺƞ=ijh-{Ax"[ў/By|bnυ٩PSH^RB ˎp\U\-ק4E[wS{U5|<3`MRb(\AϣI)pCuυ$#:4-WB9 I餐hur&B("2G-{G@K`SMQ!U^hvXf"H8V$fy0@,| 1å~nٷ%vjy r yfЮa}7Wvex;ڽ&m|Utŏ!F]_10(cƕJ!Ĺ@tVP@P(Xkjx@>Tէ:W^9i 灁2wڛ܎N)Llzy Ʌ=I—p:ԵЮ#-mPJ'UNqs*HoQԪ?~%+;sc?spÊ L+|P) ƪPYxڅ1m)(DCjA[HG';:)ZttR4ttud3?)F`p=H1\k|DRRanS18-H\dՙޑ`Ui/T#gYԣ.5Q+nA6yl:/u!٦A%XGZ 띲RsL g0 xt33J@(u\df#9,&' <d#vp9y&×\Jqnr0G3SA06`l2̕b5K]Xm%r5Ɏ͌!VkRR tz|,5:Y4IΌÆ?3*rᰙ\BƒgޑC*Wo <=Y]͙nС) HUXc+XoHq\beZ\46cC6tفiB,#)3IY2 V=l83v>?w(l1 G9aE!:Õr ^YB&[(@a[)6͇͇%',"Cʢ H\SŔĥ8Jo6k%aٟפ\^Tq6~xnF3b\2k4[+`!@ثsX=i`5&D#A<ѵgr%)+ VUK| gFYΚW&9l<3/ޕ?Ӓ،o:5K+DEaH9bJ-JrAೲ$bJx8|8lv$n^RmU׺V:ZO1ln;@vl_QkPkN6tmq륤K }Z6N~hm@;%<~J͓.CI6&cz;Ҽ7}FzTw\5L䊍H !"=5՘* ;6xNيB&lVZrEFդȁ3RzpM^QQf~5ּ,ryhݹ6A=/_Z=a kvLzl~B98z`; ^ef8^=q6i>ҿCocTd<2}[[h-/̓u Sv'Ml MLXa3o55YY#))w>GP:[Sd P9tYrzJrnP#fDAؤYF!$s34w.3?ԥ㟯~ TݭׯVsⷚ.wɰhc$"|Dg DI㝫6ޤO;~>:PBoriaB21:U m1SZ""".^QKq ,{/{;VZ]A5ϡ(FU)&Bt9X9]:jM"))hN\M:;,7STPR6BRR[nH>65vؖH+C.+!)3nŦO (Ukt11SIL 8 JjZlv\ܷm B}h E r[ZvY7#R,$bcn䤲6r&?혣|v\ǘŒ`uRs˦!]גb27 @ٳ 4b4c3K)}smA~kxCVs[֭͠e;Hr:?=S^R窍(ȎM)MOYOʨ<&/[X^JrGY^'_% ӊDͤdU0XK\I]2 wX U2XՃفpPu !ͩ؎O 'qN$c}ۅvVԋlyd|Y~gAvݵٹ8.QL`+m"./NG.Y4O͉L(zJJN9dav=䬮 Vo׫m,l9^slqc{V<}cnl;R"{Cl}ތ1[Q^dGKnl:J0}Dc6„t0xo ӏ'#(Ĝ iBa28Ռua~/a~pq88m{1ɥm;k %729(%zn|vz'9]ڠAiNsϸޖGjw՟{W{aMc35þEn% JJlLxBJ9ojd#5z%xl#Ӭvt,޽x:?⺇bmlvlǠǷֳյb;7i_GL[%6O}0P8c.dEXIRs(E7s55ee:HHո- \tzy4P@1 { \Xi^wA mh/-Js9Gbh8l81f^~'h8=Rⱑ׿[(!N}@W(a'VtiׅAJ܋.Tn*5!LY׮)8$k Eib ܌-J; Lx0)AQ$Lֈ*dS2X%0ƥ{~\m6S!|Rǵ. M? HW1w[%N'MlԸb}|V.W$ EJm5&b .ԬmJpR*Ѱ08@ od=XLP# уXFK%:AC375)u`9H̼6#\s_E!^ע1\IC29V@hkI1Q-(N+ z|Z7^?q~MQY" f ?a- ̲1Js`}7r1#?GT6,dD& MS2 YLLs\x)7g4t);L>e"6l)]C@1f2jͺsUn`rqyc3~ם]ޡJN=OR Z'1|& HF>D/Rkz `m(挒YSf BRIGԲR/XV /7V ֆ5gkڛ1}7X+1K S|&yK2B!&K|'V@& < s^h~-==9z:DcCy#Rg>OP5 .(LƻFW:.>00!Z? yjc` Җ78C!\w؅Yr,h.2S^0COE9@ k^Z u!`+*-$:ʒ2CE'H:1#W=l8qUϡsCju}9iN JgT(G9=rG_8bn$B5@$dʖrirT]H"wzHM/qo9R!& JMVkkhŗ 'IIlp¦1D.8 cv!БK5 N|BZ), ayb,߳Nyl年xMADHX.(*sI!UP{tP[B`dx/_{kfJ)|P *(84T1p*g9H “Fھ5ņ4dib@H5HqUU*`&Xǐıs6'?IזfIgjpY2CxPj$P3l![\v;":2uZWs aS囹)BEL%6Ys]I86c#tAzܥy0`l*fsG~M﫾>Oζ2o610{}k Pn bW*6]`.C-'WNj5/r9< FX(L.%&,5GN%dWeJQaKl&+R5Zcw?,4g⟝O+{Kn=i7W0~Y~ߩ8ɽ~}^ `ސ,ތlŷG 8w?>Jo?%;^3tqnKo}F8e2a3%4q7H! f5;]ݾ{?ded z@9#X(X@`ku٨\J2F㥖:q֑.`N(M|;I}P\|9Ykm[L8Xt1hz-dr%$S C1kىIQo6@Pp:X=O:u"="[tsknXWAe=5 '>[IN'r֛UZ"JJo7x$0#bG8 f揞„M0-k }>ee3ELH5p^bqK_*"mR"QQ!2b f*kZ-*rxs$Yng:[|ʃ씞s򠹞d/HwDjqѳrx?"#̉0YNHb2\%\\3"f>IʜXmM2F9`}hZM. 4A2]B|- ;q_O⥯Wa r$ͅ,`/efɒf97,h{A1{:k{h=X/frOա|cs4]^|F۾9}R~g$}xF?.8?lOŗ߽G@Vue[2__sD{rsueC\Cu1c?sHXsS$%#FHc#M޼\]%V@ݗPkGHn"/$>=fӻGkn`rNWT0Jq6{c dwwW׋Q>j%ғpDU^b=IT0$?rj)aJ1Вo S̢Ucǿ6"{76|v|F?O^} \[TI&&y& aW m+kLXs;c[h+c|cxkO`۵w4*p ]ryNOrBL{'0{‘‘?~I9eD"l" NUE.??zѫ?-ulЪ&_<z)F7]stxZ`ǀ׵azBWO@HF#'o2F,DULw.0ՈEkmX|h{8jhqhA?TI"0ԛ+/kVq#"!!b#jBS '6r\!m^tT W+C8>YW 67r"WH+r2i Z}{#W+I_.WHC#Waӳ!V'=U;\)vZ#W(-sKX bAmzJ91#`j+eD"WH˕rR\lPwК/GcpkuՕ ir9T. ^7LDD%~K6MKsx8ȘҘ(3;2F(AQ6>ތ*L!4[iuFJ˃LP'ܯ)-+]&UGd?6~Y2?MЕJh:)x}]uv(^>w)3!m|VNXm|ފWd:_\em٧"/eYʼR%E(~{)L (Ԫru_&j\6Vh=/tLt W, *XdƩLm\2LTV֔VMT mKsX$V6$e$bĩ5$vol)(p/F!ؽe 1 p+"+ĥBZ )\ R$YNO+,BZN]+ r5HRQ\!g2J ;HDʕ#`) "j\! pcOk]roV9&&xQ`J FWS_ iAz1r6=sڻ\Vﵫv3Sh:َR;v[rutKKjM56#vklƟ'73JjbEƊ4ibb0yT'Pl~U=߮fms!Moŧzs*ff`TʤI+RZ؂y[j uւ:6Mb$Lv6<#1S#1B(#)#rVp~vd[æbN:*bGb|I}Ť|9w8gWam|Et9JS]_Ѵ[ _:?Zܨ0Z[l16\xc!0bHR,bbQ 'q4! jasCg?=ePhkNi%Jj'Js~ݼl%4.YbI[-EYx=:{VR*D^H ydImv?Lg3$MKRYjrMuS!lWm$\#wc8KFlk-{u7DoUQV?MEUҺ»J8m8;hE1BD͆; j2dj^ͼ kL<`t݀[MYc] '=wkKѼ`oާˡR8y-<]ש?:&euRDXm哕2)BI`Q.C 9HOQFgWcokl ɥ~b`g]m7I]o#8WƷaVZúG8϶(" Íkn K!ݦ|}xοǺ}a^immi`Vy4pZ}90"9㗚$6(#8-[ZHÕ:TUt#n4Z>$Olpd?=^}+nUdcBR4gjCA@}G$mQNh렯oNR/fCp5xvǞҜvj54Yb|mKx-Z0o<⋇Һaz5 6Gr}lG{SRd\ i97[\ BZ\+$ _Tx$WlFWk_ iPZ"\ Pϻr\!Zo"Um=~rB\iei3,u%W+ "r^R\ɺ`{v"WHWPvagpre{$W,?;k/ri )erJ4lz1"hzVXv絫v`;JXUBDcH9Xw%Wk/r:/WHɃ\ R,2:fd$2Zvj^oN]tG` TaJ` k脛5*r|a\u6 ra671 L(oaUcivJ\0VP\pZJJ}+uBJjr%̧FV q7ry (5A(W\'B`ux#WH˨r\i"36rWzwڸ.WHYQj8r^\x$W̅FW_ iRڰv5DJj=+XWkZuR`] ^B?l,{wdhXgъ“-J:)$Wl?rJ"WHkr WC+h-K;4&5X&Tכ1"H od1"@˴p]Kdz28!8vyZ s]'ך1k&=KYNϢz8 LҪymZ,-H@VRYoLa5ShHɃ)8{s7o삢v2/?..Ѵ}@mYß-]Jׯ>>6!@oUA*ӆҲ4"_\? e6_/NѸKZכ%ew::g,>?D+#u[xõbnWZEB؞ gk`>S5Yxz?s ?xS̔g-Bzog;¸#?q+^|\X>>  ;n3p@O| }s7|̔ 5' bqsw}_K ux/o+1֧;I2LZ+a YYT.EHU&dvI%iB[|/mv3awe-~ r8/kj$Q6W%ƋjU|$$ѿ:F%,6ԭzCI>DP(1y.w6eRQh4,4/`dliF#wwBCzӴ˥_(1*XGxv5OҚZ2LFњP)ZVD(jϐN%c04JfԭHjhIQڠcS9ή!b5ksIJ;߬RxST!D[*DsT#b!YLf cSID34fENYV&,j-'"|h`6ѨO.ZF,G@똔2vڰkߟx(QѤ C(R24ޅr@cVYM[͓mv6 !}tB (y"iOus\o>P*[YDŽ(O`ɘȺU>(x(gU^ԌS2歕9QJ2ɂM s e"%d~zs*Ô| K(%b$Um(h3GZGWHV 3_eAch#JK|S.eRHE}i!1!im+UrxE1`1dǤ=s)Ȍ>4J4N6jڗ%`w"ᨌϞDLڳRȦB@S(/٥$ِ#{L/ޅ*1 rJ''`b|ZCu\qLOJԐIX<ĪuɕޕXɠ-]@8Bhcn5i ,& Y'eеZelc]Q?SI9*xYאU6Qz$ic!R}c PVC֗]((CJ@jrcSm'-ZpUTiXTm dFfC2"Q6p.H#`AV,L* I[NPN $_ lY!:@@=5(!Ȯ܁pI +Ԝ 5 e2aͷ6Q)j""*&"HLsجD y>TBݚ xmBRVLtqiyT/ɓG%EAVH!38hG_}`QI,TGW>AՈXżۊjlPl$S`U0~Hv?< ͻ< C24}%0+t-A@F4F|u)=h A7i`uj З)4 !pjI7P@R@"ePExt͔`m5Igkǂ@K^/B!Kkh N-!XXj59ҷd,j3$e#Krqصl:L!#E}4 CV@;8 jZ,*&a!$ D%#@ A169mJ?@}++u5 h%A 2u-Ei*5&mVD賂*6 2I IX@jT_j:hٳ)&![j=%zv+{xyן紜i׋ys n*- n=Af&'f5z-]? M"'4QGmERV^+(U˄$RV#. bL , = ArRXdk80)Qw 5ɡm 9 ]N(7"fh-8(tR"e]gd*2(eBAj0F嘐-HOOdPJ9Hqؾ?nڬ׋ae V(g XWTI1& - wH.* Q CߡPE5FHmR5uT/Fb6;L ;AN l?V`L{|$0GI-VPֳ_}s 8nr96igZ.tZ46"Eݫ`@o6z\djN7T?y!9߷6X3mۧ{[/G ~"~iغ9<>BvxW7\C[7a][ gawҟ(yb䟟> 'k ᪃quF>w'PGi%;^HwQN v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'ul~dN =o\Z/wzNnAb'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v\' (8Wq >%:B.a'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v:<>$'r}'@Y "@ N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@z9NGz~}޷V vZz~WW0K'1.wc\B~ K@ykظOl\`!-r>sqX"\庇%u/ty/Ve}KTޞo#O?hͮE3OQֹl#*&+|hןYZ>phWwO_w'g x{}4z=O~99zLm'п_a;A; EB-9bPBo)J_k",2H+V;ʨߨYۘ1.Du 7L]}}}:p*~ݫ~vBaB%^-ƍ<{guB`C7jQng(QܓP1yN/V{?NoVeVCo=': wys{o3xj;w_-cޕ.G迩'{d}8_]=\﹑*3bB*vGԷpK2FA o*lU&Sܔj/S̭~a8=.[2{`{ }ۼ bz;wwu?{F~¥q۷TR9u$9ŕDɤ,:=!)2%*[ _7R5m~N %=}??d\}^セǮ#8O/na?rwU\[>H`u@Im|U Qo߾lN{+j/ߞԢq4xbkr \\Io v[ 芢n 32_ਮ*׬1|!f'qs6#Bw5g*][A0z@\ў9p #<Ǚnp8gALQPJtxB@h3[ni&4& ){Xgwh3g.Х]x؍t/^ >멡5<̿PTu&yB+37«^뵣Vxljt+ff'k̏٘dȂ4H#Oܔdd604 7v$ "ZYt!s4Ǣ$1::iUdڂ\D. ϣQ)mO6#g?,`ӿ1 @ܧ/r^NIMU(V~U8~IKL|̀zkD š<2ѠW&p^ d5ME1L܌LIzKTr,|6h:> ] @KؔfݟD]^ufLv}+8Qm%ؿDVjT*a3Z!殶 }%{6h8 GK*1]xn(R)jnN <<6rC`_&!{i.G.P{5.9DJRird붡-jb44ߨ^ӷ7WǚYMUG+%HOಯ{ WYoEw8}zqCP9d.LF'  xmdZ{H&"Z@)Za& ҢoC^iF΁^!)E L]Aw 'e­6ΐ (] Bx_2i;oLѰ 4,*_W`?q! #HjmRKg8ih1%i/ dRȜıyl@<6˱-x'>x4B{/V= O/J=܁EY0BWG"R@PG,d,nc6KY_m~^r'Ͻ }KX^No˿*n;Vg[0a@gcQa.F0l2=\ ~rf=n5b`55vRڃ?X󭟲ٷ/E"7AFłȌiRRT&uj:I"!&jfcU \*t6[%;Z`3ӠAL K(r[STS>]c*程Lu49M<%b9%t#6a%z >t~؞!W~ݰtѾy#0ɯ/ABƙd0Y,)M5=1}GͫS(WΩz"FsyK>)WQdށc@\Gi:J/r@dH/OŕIZ:-L2,,*PWE ҌeRF)Z͋Q "{5'Gx/dc>%32QeAj Ud:K_ё ^ͼF%ŕ* %VPz.g]ͳ"{?Πۢ.NA6B;])=#/o荞`)B$f*vpό7ȓcUf~^<=(>`<2]aKbNY [< cK%P[yC N/,a E>d)xtwe0+LWPҷ{Y6+kۉ׻Řܶ?2{;ah.tB9i9T,]G$Bo(zKp^R]:}K"WI2 !FݔJغV+rsV5f77wsxQsEcStb,ȲR ޲zE?ص5e:̠u VI._J3n jvK-|7wݜz_D$m^v'+=u?7+hF篣buN{&}ےoF2ɖiwf~ef^{?TNg>dϓqSWL]ͳ75Wzh|BO:IJӘSEmS Йkhm{OݙjY%MwJ.ҝRkл3U*3ݙ^cWdUfo \(-&v?HA$(`t-+eg=dYiq |/oWh4MCZv0CaVr/.iѿ:omJ} AAaK%qiJ uY`9(,!' *|DZo{v鐭m<.~!H R%clQA1rX` f*V9 (J=?Z8`#pHc):Ȁ>&ZJnD&<\JT &Z"ZM@ % s/v@`7 6Q=Z7jۻ'5gxE%\ME0d5BI#PPdпD1/!OCR>YZ>IfAԐe@Y-%hn\DͥEpeDWU)W^i-RHk:b/ Lb4(Α+C}$QUi݆9s)4/?ٜv;R/],~[8z˹cە}hIB9)xlح7eDAP:/29s-dOXgA6zU~8n?O-_eNzG@axPhS 5a btYz@60kֽLm+ݭw5nr_Ϗ}޷}7ןOWJ^h-!`BZF1S;K0>+U%Q$BC$D}֎=mQ3Cmm 6k-ZYM{>Rl45 BQ&6D{p#bĜJ.B/J9dF ]w%D|m4b~j! vsE$Hx]BUFo'? uZp`hQ?Z[vUÉlZAB5>/% =$MtJRı<$}L4Q#@ X G1b%fWkL@"]+(NDt1)J!::Pyq`w3R nxdV,b/g۾y0W֏X׳+毫rۻK>o9-_+mH@}/q`^c`>eF_\$eQ1H Q3=_W1ܼ}\E8d4)Z_ ۫oMޅa/ ՏxHnHƝ^Q.qc?i`yO)hQ9ž飾Y{ļ3Kyr:k ٚ?g=-e}fXt T@,&}0M0Lsj]o:Ư-4+o SA9%͜ c qr};&ulEΆaT~^}[(}fڛtNc)e9Va)+S]doפ >UEf]kk*92Ysh -/xC\PU`KvPOfnbZe=_ϧp:7x{Wj%[\Qo#`8~tIj1K&hʀ+Tl:6~[pw\IUO1=sby]ȜhpljUf>JAR9\<ںW&"#*E탘&(Np)9QBԮt\O=3oyJOx; ̰u2o  @!p0Jym<IRv@/q^k76o)oKFy[Z "-E)E2zb~(f!l?K# [ D 1+Cμ9`3IIXA5v[ Ohy@FY] W#A(0ڄh;`:ݴ P@8#y$L8f2Rʃpy[/5fml4ݧV"{ɲ,hp謗V;.UN1*FJSFkEF }Qf)j`Q _>p!_:d45kļ#2F (A0,*͸Jb,aG!L<qǂq/C?e?t6J%yiG D:t2U, WDap8rܑ7_hZe#7(թ11r,2Kc *G#fFjq#6' :`!g!H+L0~DmpP(O.A%Ʋ1 0*ܵO/Ɍ폓 ,$^~2&[jU5vt3 RЅSB*C]WPɊo}Q8],;\K_4, M3\vsңko|`a*4m6%C WEVXcfEݣЌ4=iKb2L7F;{+k5*5J3w`'pm:+L%쨳yj|:P7D8gkz v:迓|ɉV"YJ"]A"ɮ7N=S@KF7!:6%D?5qyu ѤkC(XK wH8Zo`"}JgfO YWa%~n]yM{@yd/>}l ΄d*Yt2%go˖\W&~aa)ǭYJW$?dAZ6)jص|odZBf/o;KVyM5mH IY>Lq׍)L"KEפ RjR!rm|ST>NGyeu^1F: lp6HFZ{HAsE%Kv%lH{_{I]sua{Gcպu=On|yʛ^9>)K͔.Ols& ,:aQz1޻X$GHXX(NoӸ +tF {%|RQ-BGT`B<UX-zg՘IDk1h Hh}lkgp/dQoOD//Jf6q؛9(BY*#FkN"ڦ11nw<.' QCvY$P FXSa?ZިF eQ4%<3Ѝ7F%>4ݓ4y\W櫛_S ONTs_㚵$,lvTF*BJ a%rU.`$NId-vQ(֢[ J6'8>jK ^Ȍ ^`D5rFdְ5 EXȏXTM4IK7on/Anz^w<XM4 #A,&ӏ_DPTD"ytXT`gVpQq֔Hb iiEN[ВĥmwqǞax{jQ:8#W*8+yFQ8=w~|,G,*>M*@J.7!*_%?)&'4T A ׃Vs_`?^I^Q؎ܹ8ێZwKm /SySmBwtہ/^5-R[c4nnӦ*m6 >.Jɟ^!xtZ`#Džqj=*՞^6]z4 Ɣ \%r;J p#\=!\t5Q..EkIyS~ ŷ!(!ᴃh?~޽|m?T"%snrUԆl,˽pruQD,rŐ9!2RÑT~.h9ۂ`p2]% MpCy!{?<&!4OWI6r~;=3)! U>0L4!g6X8ÍX%U+x)uXsWy]\׺?N6o/Nޕ٤ s;31pys炃(⼺| շk=Hog< 0~B0(Y#Lv*NeS6m_9z7~SJ\YT'iO>UT-x92i?|kտb@TY<Ę4ΪyOSwymg?l/7A+mYO`.wW\o!<Ϋҁ-+$T΂OAsLMѕir ,r^xy#C1ﰡFBTPK Aq ''JD%T8 y8Lᤓp RN,s %0li95rL >YP2[olݑp#3 W KoI_1 >R(&LtJ"!ʑ;WVQ)A7o+cާaVdDFSt\sE`R&-H eoE5#o$E^C-7?>cI>p|,٧Ƣf1H(Vgx34ǀ-k٬r@LF@Bn(jD%T[(4H]eU=W6"g1*jĺ]ԈtPQ#ҍF*zT[Ј˹`: V1NO2o2ɐF(SķOKI<IR:IPDRBHψZcO9V6FlM-teB5$s/oy 詑o~RH/8N%| j#ؿ-YZ\bvDalD4@ A,L5e=μδǙs8vgNLQ"/i""6&+WZ`xP^@HVRn6jr]'0FDuBQ @YH %qY.K{\Lk<ۙF7opB"N 1 ϛfPFjPfZkg«V Ն*` ))4\IPXP8DH *Ez { KeǫU/;ds LHGO!.)% -ditT' ԓH6ZgY&C7^\Nĩq!BA[:Qf"H ZؖlkX~Z*wN?pS;{7wy)A`fO,fo~e< Iju4@I!9@6&#cZY<׈PW̛)8Á(`~u4sᮚXnHuHm6e!!I"Z_<$A}Κfz9^cĻ9Ե ݡ {1kiE5  BJ(eVQzd$")7 ºg(Փd"<9-u% 1D"eP#!F[Fst: 汾T|ҹ͒ ROB>6DԷpkr\]u} .GfqI! + ,p^[T.2ArKTIS2FGMpg ǹD,H )Xڮؚ8'e tHS804[5<.&DKq0bNGg$Gs=kc`>D_1CŠ7!Tl=dVX#Rs(@vF Y H58+*P<[ʉQ꽎uIf(%G)p!d$ u^I44Ǽױ,\s7\F`ʞΠL@!ŲԦjB!>uBF@Hc !,\4P Ȩ9=*} {%ӗ'K_BBIQO AVDxŌHLh \H&APOFh+PpXb#@wQ ڠPFɨ`998Z & Q;+9ܶ0\`r~␌߁:^٧2tVn3AأBG6YsB^P,/%sV'vR>YgORQQќ*n={ k*(u|WiҚ-3ގKÛs)t2D~yc^nxӇL5.&:ǃ>Lv?(91gb#J @J4r/yЁi^2k}_{+kڥ G#`@'.Ssx4*Hf$b8]XLj+j\ !M`+=j+!m,3D)gEr#`168d5-+HkNu aEA7~:&9(3,jX B{'Eb46%-E,B4 Co0P'Q,WGO|t9];(6a-ksޱWZ0v`4T7>G$﯏h!wv`T$»E/S#iI2DErkD+jiT sYd[v di4=,TGϥb؜@9FւK#wzVI(\- -uCY_/] /d~ۜXr&ӍKUޖ V#s9ތqRܨ;5AXBѢn18qqVn ~zǹ ܎o[3F!=.$?WE5 B,9.x?x@TA]N~yzյw23CP`40p0`^>O/m~bE]2u_fMgPYh)DyK|-"T¹H.y%\tzmz jI։qoee.e3Y ff;P?n_K-(*%tSקt/I< ;Szl{Έy?%7uL||NKOr;hQyfIFV=&~izS;^-'~S'LW~Lel{]nH^.pr$v߶o)ϾSï bZ߼M/Rj=6@imszCUUJCF7an E}B'  ䷖Փ.·'nnD e!M~jS"_X^| P&7pUX#}nܓ,P"pqUOUF#apV=f?_N3/Ⱦ[}aŃN~ Op w?׵_!]:r^%qQFô &b4rgOo)r%N< $YFl1Y/loeuޱRWBl7`xl̨ S!^x)14}A(B,*g{ }PY;F/@>(wD:qVZlNOK;-۪\c(HЧxTFZm%i}dZHjCPM!, 0o(O&^Fޡ$IuR+|e@H&YnCu(ds#)s9C}h!ƜEm ]E"e1Չ($.em;M~kxt|i쾬א~s6\Qkpl?#t1M_!kZOz2]}VGȃ-Ít2:+()cÛ2?.[9mm~pnv{~qu-χGjޟ0'tZu Gt\\;/ݻJpfuSK}4Sgõ`>E̥u(^n0?7h})n^MᦲhzQR#p ]H @Vڭy:s4~Udz/_{9ωt[8pĀt|ƪ)FK]`(?M5m`s9y )B$ q2ǤXn4A Upփ mOϠS(KBYڞ<1 ԀÅ ~mTw,Bq !,(W$ᴂ(6/WEb;uWIߺ&ZvkwysƞA}ԆN{ADS9S9exwm$G_Q`]-]0_V4%)kTzX(eXTTUu=$Q \ly؆Dжt)DJ^D28w9!HH2j#9v3jϫ nSKlU%6/'$_a%\igN[lޭnBeQ|P{>yNFF4:i=wib6<*3ݺzfp?HK T(M51Xz@`T9\֭N,4B 󔸅tb  H٦d> dyVM5Y7>o40f"ro_a.3CYiYtc0\e[%c>XE2h))Ps |Vks?!t֠pcp2p"&y:;dD# _9eS0vRo'-'vgb mȠώXv$o 0eV#S:*"5*UtbA¬=πY#;"#C1zzdɒDTYZ`~n\ PK϶|ڒN/z-l:dɷ(l z*@FuFb`'bZ8HaLIR4@<aXqz,2ffG5 tv@/ z>) kL:NYl}~rteJG$J" 6y`I,~.'c'g'5iq@%[ q0hBpʖHAe:Ohv<ƕh"f6ѣP=Iܞjꛜ^BidF0gάLz9Ǻ^yR* !YY)oq.PDYK֯% \Bnu5q|?;ҳw8P{椋>R胲)vmOλV;s?Zѱ%*}l:lXtYaQםk]p~pqŋZmκx8!ݠyw&>4&5tMoe 4:_jƣfN}7th n`h zce#[WɊHv̵E \&Dʆ@ r ې83`Hc@1"dxϬBZl̒ l켈Ф&Ӊ+L Us?2*հfl2 ߲0гk֜|`Nhq0] N&o;мb\$"L\(] &DFm]ĖVfP=I66A 2v̒ .9̂7Oks?b0^11kiǶQgwl VW!8@'$lR BRm3MUgBB \kUN-D[L[PM_(|Oȶ($WTN2yɉD_2u#8!cHPZ0thARY%5_%zp pѢR#,WhϑM{fM|WGG+^˵+9-:+L`\I9&F9,:gu2ˬ" )9̒q&--t+Nap>)qu7ER Lar@ց  ͈DvMZ2>?u+{|( ,ǥ'Lǣ&ZɌFA"׀ l6̥RuFx\y Y&S` h:Aђz.|^ۿҴDV؈ŜjK#M(p-5w\lx Re7KD;6*hw.q_,s62n%hGYDv>RE2EbZ(vu*uG ƿ/= Ծg7C,'ōs2gXzL aq\M~w {yh53_ttѼCA&P8m0v3٭ )9Mt釳ƍ59xvP_L3.ˁL!ߟ:W"W]yi~ځX^џ ̩#xל^Ҵ]t9)q"n~6ᴜTgr!y;YκZZQ{;u,59X""Zఉ'8h!]S.)$>k= M)~I"sK^q1Ȯ-uq2IdJ܇4pV in_W5.%&E?io}e`JKB~ 4wsT/}4x>=.q+;Aϳ?),¯D1D4( ڀm6jp=&7zE[ui/Ur:s{b6oX$mAč?)qSұ7$mhO.5#&%r+X!c9:m-jY¶׌'^2Eⱚ^b z7Jɢ5OR*6L]׋u4oWZ:Ԍݼmo b8K@'yK0 y_hJZhW=B Jn8sW޵Y&I2riKO-Isq2=&[pcb[ K.J=A5]u٬@RxIQ&H!2Z[xE"e h1td Wq .m'ݨ8NfJ%ք*Ms!V{s_gpt1\i9Vj-<~A$DO UzCgnD2|;Z&]ӄ?ĐZ!f9^;^'Lu*}!T垂QW=ճ`ɨ^ͿQ,D?^+fQoYhٿoy}ś]g`.$ecqΐ"z2\4'/ K"jV wzgHԶI(\- \]"WEJٻ_"\)\ ,z"\qw,ֳ"K+]f?ͪM$y,ns+ ǯ/SBD^a{9g1~Zxj!$Z`d߻y v{6V6+pZ\+\}ꐫcsA6/W^} U.3J~}L˰J{9Gbp9`aFX@PPRc@ȣ7"1ouvz2<4ޗ r}ތqrht_N7C:kG97My&EWp//mCݲZ~{_o]Z@cX7! l+,VjeHYQows+ OJMJBx1ر,KixGB"+JH|rijR]{)~~WR)+٭eڶP^ l5goɩSP+6jvɍ?#ی|UR &xb(ώlԹ5(*/W"yFyb@:&ef΁]bPZQ.EFDbĹ*axMH Ŧ?N@[8&>ua/y|Za?|&qY}NѱDm#I$ A-lۏEPfq-2I+)J45Fdp=5=]S]]Ujl1{LT6has,Fg):$fɉۉ6OkX1*=XO+=oe?SFH{gR M% OR*a#FHd։1tQhA40I”mQjͅLe +nF) s2.dὥd#v&nHg_ Z,[Z՗XF^q:cmɌ>iZ/c1@gdt1k[*Ogm[CnX)fu\$Ϙ78H8z3]p&gB/p&(sJ,FI$dF #5!%1:MP<2h;řQ@ٜFy>r pdGJG3Ѐ,E#s83qv. ]arU 6yPBnq>EXO U/ٷvUvzܡ%3UiӝՖ(&^TY<=@@6Dit@DY0",23=@كFH@ Y%Fg)d)\DёlQ $:j32Xbںȁ S +*%Tʒ;f-HBՙ 0HwdeB"fٳc?n{hyz3ZC0ڋb;)=ε.2Co~eWY*%DLXVN1Lؘf$2f.nfսe^HYġmD V9MLOU k9KMQ-w}&<8Y4g-WOe̱ʇl* TbY-ȏ>zmhv4DE?25N d9QIdRE"2DBx,0+&F,xEErL9 N-ӘٍRzr1SiIH2&J=uF9\t[Ի-/V]OوBjY6"*2( GDћB]tn*1ԎE,5%<:t^JM*KpxXW:gOc2hJ4E ]bb¬@@0gpzn.{Ή^v&|m+߆Pr:LH u> '3eT283V@nуީ@`'FUR :'#a8->Ϟe%svЫ$#L kz)B&YAzcsm@~ϡk?CpEB[mhx8w3 S'ԩ$/QWԩEZ:Hi:LjKvudKKwJ;Bٚ]$uR*WUJ1]Fit生mI*ے8u%qlK1fd`9$KUf.jQ2p* #cuw+k'wBth3TWJL$i]J%Ip#Mr Y,W)D8 ]Re m%Fu ]DD4]t.FO cLU3LAj,E2pO'[Bd X1AY,=0Æ6WO¿A B&H%1ĸ@<g-SҢ5JE}^ד'&_A[l*s r Tg[8"9u8S^Ig_mq grSmK 礳AͻGLꏫ]T1 0+rR> H ( BC~W > !+PNUYPC"J Y{CW.=!~ 4B<~awo.yצ b7Etyu~*:ZļTRZS1[)MFQ0!O|ыe Wy|J>U>u>Yl٠ ƃDJp)iqkf̈́3"!^ =ʜc-I kðV#m}jF3zGer8 ߲͋} H.@* w?}(K$U{O UN1[i!lH9 ]܅޻(u~sx Fbulyy_fnU|,mb{7M]jUa85"1ɐQaGPsIdx0s SJKhjgR%XZ`x1<< Y()? OAj#Aɶ0\B) <@4HʘvKb<]n pQx f8Ξrx)0ȑ)"c0) J(,@$擕l:r2`EyδZ{އ-:/80KE]3- 5,Bv0EF լTr3sdzȒ+4+IEm`9dsdjD2h\`ƿߌv t2x'}"mk#NFȐ|h6u4]N}W-2`IQv}t^}z7 ~|3֑oұCm]ܫ cңeMLz$:LoD0ѿt% Zz&zm1mhòmYט~pi4}-}{;ܻ^Խ2Jy};Znd|k ghD.>oJB{r55F*h8ae\F_lMtMbl=a}[v:r(BK<l6jT{\i?|?m1O8?F;L=/ozbZ:wjv56{z4f R@΂L_)iḲE֔&l j7}I8(eOOR۪DQӒ>' 禟ܴ2v_%.-TgIiB:A*(*ǂqpRG=)ݿdGFe/ld0\h'cvI˕r$Ąڬ3ENsQR8b֦)$u)E1@"%.cF9f /疺.љ8{\אhj>Ysƺm-Z=:^xk2u6쎮Ux'SqHwzUElYvJ>[\0M;ou41n w۹yP{g{Uޯ]WwtܤsywYM/[ҦN=cÐ^֡nL\;e屸,4>.w+UCWTJ~ۓnIuBuDߝk*cy^uNˊl@eO@m{?хe˃6}3R;F1|st1#8Iuns-3/tt4z˔SJ=tyْt|ei֟}(6@;Pg0FmWq-,cPcdZWYHiFԿ߿{Y4y+~4c?Oo$e2!el 4i呥]BŧIiת9E-a$zV/>Yq[7͟JԐ>(k_AO+Itm6MbW%k+7Xu}3^`/?ӛ)'e2{#jofvr3]J YKp{\qbyX-f듋&z"H/+_eL[ yXni;CQ(n?lNx \>Z'&ob* 9k/;mT1/>g筍5|Ѝ9w^~'1۱`oUvd_@"&@{HNM)نT!yjmU:BD|Z wFv& $.G :a4<*6.M-?i»X ۆ K``Q0fFcKΤа/4hp4*tn;:3f5-t^A&]tTd`.6 R!ifCUdkm=(,@@ m #"L4wOl_*|𭺯A<HG3U:t=UJ8^ǧt)늼oDZ^.ݚ15_n0=Mn\F,*7M6YMJ:  ]L:<~KAFaֱ JW;3aIu+7T\oӠyvHv5eQڃz*YFD$ &DBcv>D@ 6)+>;uy3ֹCW'9:cU}|'9WG)VhtjګJ}/ɻg7诿~M5mfybH]h/!|lK,-@#79ô-NYgem,uRؿ ygrzI?Y;Ҿ.=z2>;6WT(C`R༎& !LXEvޣhRϾTSYۀ& òlB$]$P^E*&$CAŶM 3W"(-K@ M?Y?$e,$Nl2FM O9wXCl~ъڊ[I+'1%*߾4Imsm<^ ?D3)Kᤓi瀽Ew5vƖjod\c'٧dz탿8d6cPA:+k*a!$6 ?Ve}IYL!"Cu0)cTZU֑Sr+ ^2 E9#v< soPw j FM [!죢"[`JXAkIXPk6_kX2[؋&cMʂc1SNkFx;*0Oߊ?vED݀"*ER(b@AU:+a(N&0F8-(SSD ^ a#62NT8S$ffsEgOĞI`ג1"6#vD6TGj }u6qQƸ\p&/,T 'm .#tcbLd/)QTń͸cOpy{#@X}2W_MIpCe?jоd%U*EΛ;H.vXlPe}((%P NT~.ִF6_AR"PFPw|pYQ=iնF+㜬JAD>jmMldu(&dILuk9wR\>l-rxJ~ 9+' }))Tv&>袋IȺ/ M9 ϸmU1?|糚WwEmISNמ^6r]W% PNJjS%CuqǞq$/ZE[xKzΡF#Ӥ W 8r"4Cե蠄tO0rh`b٧ T]:NK1ʓP *`YmeV kbEG ̕BvI> 5XxT d "ªfVzWnYVܹaǿ# ḱ/E-~ާN`Y Lf5_fy\y.&|A(q駳bzK˷䏣P,m$3ţ&TTBԗ#ѻeflUؤSsB9z='|b@Hm^i\0䛳>U_QzG NwdAk8G~r=Ew ?3GD t !AuHE'A0l.iZ٦k4jzCkNt8ĄBf=]:/2ö$~0ԳN/zS^曅Sg::C%v v=7 ^z z'PO")ǻksi~}?bjPWJ*8ƾ].]r&:} ى]M0GF =?hz:_x4kL'?H#ewʀzVZ |~2^>$QI •I:`YD_&lN}$Vy>=Y!{ҹ;C@ opc˄S7n>={?{/2Ao㖀$uM]SjzK>hZ.exYW*0^E}d.vu`<: CBHCSt>d:w‹INH+*H]EWhrPXogZV_`unQm Tl9PRt* )PެQk7=umN|&-E.C%H4A& ( ]R_,br(L$ k=NEKE $EI9{/Qy8`J`NJ6NlF/sN'iO+_7RcYB .Fb`J }A4JrbBCzm> 쑤 ]>Pkg+L f%ejED)QWN= j7RhTjECT `b/9 +%5tiP%Y2[163d(E]Jr*l|V)>baD *^I2 c=4Sm>X~O!k_KG5$ Û2 =3 aъIVyx9@dcoj3#5KvەkZfBq:rA2U> J!XJ C7!r<3CN˃}$ >Jw<&<MvϭO7SxJ7%FU:xOj% fR]k>L560r1;/ VY)O94R1 /_,<9RY@dD5D.>(Mݨ`DQ9-6iN:(RJrI?{WF OIp2`H1n?ALʤ];Op(Q24ƀ]Nꧫ`e_'JR3#`.=)2*ATWgwxl|6G{4 .Ѕm7O ^|<+OV'/fb;s8~nx Z^]R"fm`%DLXVN1L1eȌŒF]]ʟ;\ )Kv" Yq2UWrE.~~>j6d^tB޴ QRya-sSzWƄ1ftu",Af4C v $bQEaQ+bR)[3_[!%tbb(Ph XZkcuJ1fK9#5G)rUre=G0 871I.OgWrY% HJek[7p>]@g_-m+@g+qr4Fz 2L`R^lty(R h= J)X1XcRtM)\@$x())/yT*Paze,gh`K6)ͼM^ɀ.*g&kz+V5.G?i s&lBVx2 e yӶl 3mڄɗLp'GQp,PXX^uxhI2Ӑ1%- d^lsaf| 䣶UlWԔ%CQjSP哕ouk6:d$a5y*6 ar%qzf+K?ЌXJ0` Cюj3PPC բ,OtsGԤi4]JLV $ dEH&OD6zZ4xN: 4h5Ɇ@oK"n^GDGrXU wMcC- SBbS6,Gsr5riw ]մ];mx;;tpwX:a7e5?18E$&38H\c@dpst!eH:ϧ|j( ҏ3=mݯC?rӇ/w [ˀk-y.\$ hT6xC"1gmB-hb\Cz4 0pM{Q$#鯏!z99lx'H[9`T3я`0j L,ɇ=tͽZl嶺k#]m<*6HL2dTXmzf\1ezڙi x-Zh a4hyj p^u&L Z3'Ol9Ȕ 1tR fa 瓕l:r2`EyjZ?RVYoC ^gww H CJx/ 5| d1  79GN,BBTdY)O8r Ԉdf \ 1RCDZڛY !TNzꋁUD@xg˘-ƌ Yy#*,E+\`M)FAWf`A%167sD3~wz߭xQ[xiRk?o,7;+o-A|qЭP-ټ%{-\} o,&_qU8l^Ҏ*3R -j&G?äuҮL=LJr.I/Z $YK)ou;nsx;)~stWAquk4kz[o l_iŬk~6tN~^~~] 0YvtuՋZ]\;Wbi| &?d >XhM 7 Zԋ A'J\-%+FFNjz`Ш Dd6p5Z{"\up*Mdzstڴݏ0@{?҅U˃_KX$vlvcRkXۜ01A˸0F ` 2%1D]17[g>>#mvb@w&6*%<.ji]Aw蝔AH ?N/g!nyZDurxX1p OGƐpU4$׶^mXTڏ5 V6T>3GouJc#+QZӶIm]ql kl @ A[mx@ari(l{{z+23 Yr Y|{=y85tgO|6yg{XOgBH&/WWR#-jubn<9zvD{9O7~u"v~mg8 1n}z7O3Y'?o_"ԭ";٬Mٙ|cV)ܿOO cy5Q/O񨍄Z\9k "]XP'n8߾w" <4F(DjBhp>=-++F^_'J(}} ]+f*EpۮŞ}Έok/hkOb|,@{`vny7oxr7>َ.颷!,oŘ!׃iH<ӂ7;W[JQ[CMW|O>3ol;߱ru1.z=4Awxq:&hMpQwlMH9jt]oC<Պ@EYcy3Q2/PIZjN2zMGhQD]lRlcccR1LGS N4d9<ij@г3j=+(͙׸_M9C%8M,c ా\~|4>78?o L7AΓ\G48QMԪq('(FbPF~jX bQJzI%d3YlqzjP&hmUNA5 i  ϕyXfS~} vM^.۞M$hMZnsE>Jl5m~^]]_iXłBK%P3!Vt6_MzCܘQ(z@_]zZ%^qfpn濵ƈ! 9uvc1t-tսܻt}G/&^(tz &9Ul[+<9QM#]Th|Sq=f x{_.v:[t>lu"Ox ӕvˇe3klW˲GAε qC򲌐CfwTfT",*IגWӌOܫ+2yC#kt!L'|g-q,QrS)'MM/+gShIh MNII&DQ5N:yŁaL@Wm *ʘtCS@ڈ01mqu49_~PS_t~Mni~7#sj@&E5FϤ<^&9J2&F#t`IVH/Bh16Jq꜈:UcI' 4Zz׵m4c FzRvب6w.VZ:ICG2UFch\ڑYh* z7ʴyx\ry~y։T(Vrr6a5uA;H+T &ߵIuʙ.J?.C6`/J@(PM.ZZH_Gc;i1i]LΏL|bfC#jJ~v&y!d4F 5u@2t4ҭ 9fnTvƺox!PjEGHX&Js#xMۙVDT'<MΗy2ӏe%soS]1t]lBtDVARA F)0J:6I$F81Q^ a#6 ;$N9ӴM4L@1v}BN[ F&hrLo5uC\\sis4+y&w E_X9ُ]tIۡKAC1b6pQ1z!0 01rkp1fp{< eTE?#W  ֪p,WR ±Zr/*wՂp)gc  զ\Z}JV|q i  P H\ZerPq5A\JHY XբWR [qJj!FWq5LQ֟WCTf+5WM/O]+l +{jjZsBU\MWw(`å)Ϸ%Lf?쬄+X߯t*=ٿ Ϯ/o6mn.q1s$/~ pJ # ~u+ S/4 Tc.Ӭ1* H)?7Mͭgj#tbXqv.FmXzۢV4f]#]؁P?->vW>!6f\ٓB~~{quEms6hk31f:6VpmZDχH',)dq IŌY~*uO)4VPH*_ HُaYW˒VFX+gez(WWR *(W,Y! X`*;XFtp@[]N5\ZD <+K FW H\Zs$P$-W,ZWr'BU鸂Bnةp5LUăaj}Rgg  *R[TAb^+\Z>w\J) qwe0=SrN(L3 tJg$3uu+v]0ts}6Ws~k~+V|,31V[ue7~京!/ I?Щͼ ͵J8zU]};6])3OAQ:u\8faj퉼a*siV\=饶 lAb\UqjOUeJ+&+b'!QR$ }f¡ ڽ{\BYLMEg쁑HQ YcP+ 3/3FFH# rI8,D)053*mu  \`[ Xv՞0^V\MWXKZ!˅bpjXuWĕYlB&9y-f!ԂйUʺ;E\YQr<\`b1"xT@(r:_R >}qarM1AV|bX箦+o,\yUX-Wg+RBW\MW^+\\SLT/;XW ܁MzWk85'3L;0>+7WM/pł/W,A)b(s.bQp&=Sj.L34hn B8i7q?KZz3 2S6 0H<ʂ^qzY.eVJ';AE;bpr*W֟*zwJi*&+%`,g(f*;XSSĕA*W$%˕\ZpU)*ĝpkU$Q׿w.Wl2K:H/R^%\mJ?ݤnWM!;:.S;ꐩMzIavcRvnv/7t}ogzȱ_iy"#_H>,AyI<++KԞg1=vIj0]CsxvzEC^~OM>MM| i3Ƣ&^mGh)TƀnnmsϯV=aRWc^ߜ7ݼ;_TrK0g7=tn";6CCۿVFt-uܙh_ܝml͵nv=Vf|~N~Ѹh؛Vpm\Vh;v6PrGo⒮or\BW?rdztb/娫Rj Go](o#>*b F=7]W?gL0.܇*]=uI9ĝjfK6c+ ]@;:ݩTT|08:݁/8,HUh?v(- MD1zs{И_so0 QK h1pRD@\PQE%[] q95,2@ḵ@ +G!,GmCWs?0o,wF^"]y?ޟ}'{9 ezt|$k SzM}9݄[̔=qYGjl|G_ɪO܉߆o>.,@ED n8 E >:}7?ݗtNjU>ў/ۤ?˱%mC%]|059:ϨʍT>$򅒱x+߼w{;HYp a7NoZ}ݏJnN/k#\RѶ3ecrFW쳣Vy}gf:l)4jLBJBnN;N9W]J>~ƎTgtO bۍ/Zn]Us)Vαuc7SW89YtDkc sRkPTȁbkdr-H6hFJM.{OExKHXݹR6SDm5@&g2klkdSSLj92T#b!;Lf c#{Β-b/9%干}"6Z}xoW:d61k(Wb݃6\gce6&]J'M9w Fa?@`B. YL%vN=]Q~B# PA\R}S [P^j*$_tB2K&@@37-V(!Ȯha@L"7CA ,ȸALAAX' @4'X!AYP.34vKuEBy>uFݚf$$XYu0wT `m06˥đPI!0Μ zN 4XSy(y3XGf2~.~+B4޶2ݙRPHq Ls$eIj lgYH( ;:Bڌdz$Fu3R1 ATݳ.RS+է\`3/ʺYHH6y6@eR":J( e Y! CcFU+czhA  ܌t;^̈KUUW̺HNoT UllʐvNDc ]iy}g⅓r~ 2ZYIX  /^m&YVSƐ%2x+{@X2]lnޞ_]=Q$7Հ ;N lfѓ2EFظ:4MvIf1hkTךB$KBq)'FCoׄ b P.z~==.iV{ұRpPaRDlwHj"* F mE4Y'%\IaLv%0ehuYځ4PCz{:T}fQ ƾ8B?߁yE#xQvk~!z7pAW F-l֣,*1"@uF111tcT,~ xCڨKǨZ054a4aoLN+qҡƚ5AF %_64d5d, Q.#TK1>dQ?MNL= !{ *uz\>@Mƛ΁ lWlЫ8vAi Vu@9 Zpi/zBLzRčt4YJFD5SQzB%a9#i5VCo; @.*  Yc6ՔWn7KC1ˡf xHAH28ĤiBeN Zf !KkwmT b!(l 92L0H5ͻ^c!.էj] YU3c4O2rbrNVRlحJy rq..Ɗ/mN?>(fqr^|t?qO i>l/Wkz}}{5i>UGg8?eW׿'u;δ{ypQggw_,<BvO֫9ڸizY%3;TD'>tf?`rڪ9('ЀX @;Pzy͋tX $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@:L\p\O-fŁV~zN g=;qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8^Q+l3Y- b@@k;ҳ8^Z@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $Nb_Y9ׅ8 wN A@/ Ǖ+qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8^˥/gchfy}~7Zݮ_o0 $Yqi,_F5>zPj'ƥ`\l҈^ՀKn)t5ڣɎkx\~v:p0 Q硫Pju\t(tԡ'e6 bj]]0ʹcR +,)V\ݢV ʻP>+g޵cٿ"a~h =3tL'hik,KjdǽVIlK%mU\H`/%/އŪ[\u2#/yCIz?7K#!YPVcgH ^|+}JOiF4iY[hZ{4 ()MEDz> +^&UJ7~F`s>Ga Gx ~zdp9jd4.gP/ΊR!1~gRb6 K:WY9O)d(yTL퐃@ѿSrBW?W׼g~-ƨF[7Z3w QJۡ+FPFтgG *ʷ͍%ߜеךl K- ^jp9eF엇*9!mqNgxpW鸸s_gUa]nr L]_+'5rU2'3\@l᥷Ef<`^'baxk0խ=x1 (%~b FemOW#e-'NWR ] !nµ+@k<5y;zB3m[DW+BNWR+%Etp/-th+DiHWZ*kۤ|w۞09Dkա;yte5tph ]!t(07IWV1d kB5 :]!Jڍ ~;tEY(əxu z@vnh,JAld@W}*Cy \im+@˨>tBvtASeAΙ{hAT>UD1 0͙!ߜpIn#r[ܗKmͦ9;v+Nmvv"!Z0l7 FK(8x?'ۉ%G=y_e)y3>ϟ?9x|yxN:02 x~ByzXݧOX\O{$膃SvIඓO28. idH5Bo>oiQkW]f9{l4]_VHj׻[:|=6Yr3?Uug𷹻ByMSLEL0).)SǤ'PM:OZ0Qy "dꘘƯS  Oc 4WQ{`4(:yV͇Í4ia<ƪ%64:uЖD 4]+{1)}yt6_֋57rkF|٭?~/u6:ן{f\:OuN,\ous<?* 7=q\2< U.PH.epAWA[hXt|/a>-AU 7:Tw+ -zKV-8 0rZNSMpp!O^h&{@ͮVXb׭8LAB٤S}?ex> .t~$~w a&Bt^Y^aѬWiMx~?}F JKQOOCJk0.֭\BNƪPXh! xT$4dݩ *اtX*>쯿G6WS6²([[Ba:(Ld3 4Iഢ dJjn^zBd8q7NɥdfoM)zAyoLb:S#u;`PJ{*WsKoESKdךյ^Mx_>%miTB{ T(SGU_us"mpfD 'sDBd! +psJn)5)SL,ъ4X婂mH2%o19hHC%ь6`eZy, KӴ2N7\=~n69/b'7_}Pq'ݜۏl@SN&6*AjU|auUp$H Ҵ7'}KYo\%HPC(qʚ r ̛(TӠZ į1f ._Sԓ F꘭gѥhrN׉J2T96frj eDžʅ+DfHIELO*lXnǗA9zI%l!P\Lee4,5"#{&W:(TNQ3#da*IЦan Ìǣb k7eڪcN*,|Tgx2+elHIpDdg96 w&,vIxȠ /:kb"ALP 1H8I;06a>EN1!1Oߔ}ˈaF#vx+Y,\;D)sx.1 1.5%e^ZIHQF4f)Z<r'xLh4W+o0#^&(jXuH}YɞHIüh:^x>`PCh -G[]4 P /:V`:^|^<>l:̇@aiW,Wm䃽.ظuzxZ+xTH: z|D\JяXvy'PH\pmjD+OFkvj-ںMk/[rlϢh3 FT}2Q/) |Ѐt7P3i(2Ԗb5cL̃$F18B2hL9!N;:D&p֨N\&.%\hxq#8qs.Z Gmii/18 ^Pxe~"x-1&;nŶew8>9R,7RtshtVI>g, Am29sL6fá8k 7getyCJVb0>bxÁ" %j"-c*s*1%Ahô:kCX!5.F$!).h؃5 Z+=ITjt<vFe}s?p5$L&ͩ`^*'3Pp&'tZtRMc*#\1$d&%LJXOZH6BP]Ej"3t& r1bM$p?ie@qȁ+0sK #!/e?Y&Xu} ~.k0*>irp2>^^4ϣdWٸ_zht!y\/ßzm4XrY?j' R]0@)Q읥Cfv>9^,WS kpQ[wp7cPꞘ*cc_䯓qI_-E%^;U)v{V)`Nٗ|#rfe+r&w /B)'.k>~\cJ}1Ԩ z<;TWn"Z>ߋ7¤KOTW kR`w5E6^͔28_nC8=#;k9ȕe:ërQbqnQ$95@H.>&mkXB7^_+D*˪0:v*QZ1U EuEU-[lφ%Z-{0_2&;,p嘓>ǢBk;L/U  1S {b_~\fO l=^=LξhnLP&VȾu+ qB^'BSʞm] rDA IgKR뤥 nŨ+r,4s^dg˗[0&V8LfgF(U9hٻ8vW|JU,>j|H] @,w/ҵ,]ڿ>Şi虱# ˘aWwS>Y\V)B!(a+6%UO[M?g;yʎY9Dv;(yM>_/m k{ip(OyCG㈭\"U zf1]&W&TQΕ$*~UqA3x& pj[Eq姍Jc9⒡Z}DXMBj<:6F͜Q"tw'BҢCNF^|K| &* ~tGܦpDI`a8;>46r` wvZhtсqp&MgLΤILQXk)dE ynd}W: Pv0t+b.\MdUWiF]t$d@#SN*r..6!@ӧK/?6}cCNӢLƯգy-/3&fl~aWRz`8_Bav"@ŀwsf(17kVBuo7Ry(QnU'[f).g)~e/]^G*%(]LuX(u9w*e)̥DF+Kq ? ?yQ~` @4'WI% 3D澶)H1{9*JMP) ƒx]"Ύy(e5X!Z@s@guKjd:9Q^^o-T߹hKI(!QJg*H*TٰJ0ff].yٝd yJS•kC5rVd0o)I'Ψ)Ԩ *j-GBK,T15MM6(ɈP#%CWJdPydG}N+WS-$}L`Sm2X#T46Q*C XUUeBA&ըc򱫒ڌ^J'lLվ_Mf.R\ Lyr\=7e&Fz?l} J4mAq".]th1V*sާSjKԖs-gS[ʹdT !lvP&dcj@URLG< &{Hɉ=x Zy[NoN.Ww*tMq&s/DBrY>q5B_)ø %鐰tQ\iIsNh4COhө9qWmB0M<gkbl2ɚҒ) Hש[S7|,tuD툫Wtq\0.["%g.ԢFsp$E[1\ U r9$n0d4s~AޝUJZ!j{B աˤtUȤ]Bbq,q/yŊ6j@OÙ% ¸f*=`2UD92ÿq_+ 9ˋ{;&_hdjH-g_m'^hbP>q:]mԼ]yUb*E.#C ƛcXPBf}Y_yd)V(U!j ֑e[l[J.aUKBT NxljR[~§a[zBz8@ZBɫSeJk|:.}׿}8AclvK@|gհ'Π@SfSݜ_mH7OP?_sm*=~oNݿgp e+*ʰ#s$DMy&Pʃ-9L 9dp%~Pzxgq"۾xcɧ_o kR6$ș.>ES@rU}-kZ_^u90́-VŠ8#cN%=|>ꨄyo[Q꓃01yqG@郶ռ/ Ad]|h=Z;ͬ2 X$G^[F$8 *PT SE:@Bmp%ltc-V[Ci<<3Gyn ٗdV8D-ұ8c16hbJ9UI:Y6m&DBF8~>T3`Hdc|3/@TmӼ{5׋|uO.#a H>2tŋjD0U -i8̲p#7o)>nŁm@~n&nhEL[;ڼ{tnNbZSHQ[1W@Q3'ʚ-6}1IgJ*<3ʴbbC6B";<߂DYMLj-Ї3k&0gZLC(׿_NE~s$ͻe?/sR#ǪW,"s(}/oڴH??'sa{α.@qbuF̈́F$F}u&9Ikm̹+PP.,毓 V&Tʣ"[_ TiJ9q$&cfUdll.19{\ 4z~}t朹wqm2\.om`=1律wlu5hz5 ujGs[M[5Kn@h\ǰE̐T۾(ogw˳G\oai>Xcl7x|{7^p{룻WV~>q^+r[| .o'ܻq[JtOݷ0wk|6"Y9^ѡWw`n37!nVC87_*G7`#bp60S)3pjX|uP3ueF7-aI]˫:{i!Ȗ T ԡWr`}}<;QNø5cL!ƶS]Za\;gr%MJ`rk,٢I.GJHʻ{e5CYڙ¼q@Nad m4ȭU ܶ.s&QَŮb FS,,C貛79$RƲ6;SOS4I+ɁD0'Nڨ#SNJycJF<`%ZT~i-o);].b Hk X-FdBMc-͜ݸ)~~崊,alOT|PQK9xxCmPܨNƚSxmv8Uf\>dTRZFVq\?빓̓g?;nSGZA6e:Ps cJdUl 8O`r>R$ F;RTtAWԤCKIWB \,2“F3gO?;`[DtԦ Xb!! (N^\P!CjMrXb _J_=K솂gv9IJE@ їhU18ydV9#~|P+]B''rYJfW3 \.Ү,3=5=UU@\1l92kVA^eF;e4spI*6۴.b TDG-,Ғ)((!R0?\XL$392mzDiS:|f_o0ٕ7/g{zJԊ@$KEFBoÜej^1L}2a8 {,E*>p`+wD'M6Vg.3]`H,R\>}q\2H(&_(D&]VR[=N'M8j[m8=p}}Ɨ!R?M W?;?O5ߤ2= ]?a+b;K~zB_dY.2m),wYDW8:mY^up!GÅ%9ijdqh7n@Ivτiy,1poTS:lr^_T-Λ˛x߭mxfv37=TǪMֻ4*ث`-qx\5V.PN>\Vs4$ˆW ]E\}W KzJ(< qk.F\%t%:q{/"^H<^9۠s Uf$ᣔ:pzyF6ssr?jnmj9A)bIȥQB͛eoj ?~zrjGՊS 8w.)K'],u9z{;vbZbuYv+oX~7~u{ 3H{`I Y>6 c o\z䡮s0=1PjQB (dP WRYSޟԢpyENRQ5!HpdB&F=MnjKONRrgZ1)=o+ ,@S&aJ;f2RʃpG<7ΖtW|:MhX:u&BDba_ q Wenh &c Kp+>zT!P孊Gʍ^yQ ]s0t hYYT_sA4`Q8l 4uV'y BhI0˳04ԉvL߱1r*c)%M˃{DJ"O 8)0_*S ^皤דl_#} >P{6Cn=SL\2|aR.gʩ&ӓ{׻N/>!jhkFHŚkE%2jPP35!gXnDF(KBc duђ( H[Fq6b2l Qj;y5 ZBwK:_E;fNבuq?__@VYʭƔi<\HQmi#c8ckYP4*kcR02(:tۨQ10!NP|th yYx7>N}Ebdq`TvM+e,KV2bJˋbIEiyQZ^EiyQZ^^0iKˋ򢴼(qEiyQJsKˋgEiyQZ^EiyQZ^N*-/Jˋ򢴼(-/ EiyQZ^EiyQM 6Q\8*dQ)أYjjs 5P9-jA)H(¿V8tJ@1VUvEO By̆O1Nvߠ5)dY+F%-r 8"G&S1G5rneZGEdQ0AvbXĦZF2>7:s>&Ni(zY9I>rIƧ+O UpoYY K>Jq@(aF}NBL?@ޏ"a)K؏B!!4R|+Ӂ?󎙷}?D^3܉Ym55sՊ!YsBd_LWڻ"5;eYكLa>5xsXFMuj1| 8wAHNXQaI%h+ZKDSN8ծad:E"y6ɃXFq(iQ.FZ Lnƴ6;J (WR?P$ώ']&?_ DD/,9mpHGr!Lt C Z{|"c&81ZyCc~{gfZCRBQJЖ> L$җ≁Aך{f_c= əN Iyd95H) 8ÍfMSSnMNw{=`3m z9>FE]*cOjFP}4X!?-x xηfA@jGtw{WB} 5 5(D0SgmNƞUy }=Ftv qﷳu aY>@l[܅nM|rKO24*6hsꛚ NLoZw=$ ZgnTIU̬2ާu-aAt~9O^Sy֯U7#'%Yד+{SZy|})9"Ӹw}1lmo4?M&W{( {ސq '㦹mfw{cٛA:Lgn3LWF? `|3 n3ƣ=OYn1o1E7-&Ha߽0x4 tY1C9gW?Nr3O6o<۴-i`mqAi}iʚP%%p)9QBnz/׭Y:D df:j _  >CDa&x&"5-F%`(zj.F;Z4{Ӛ: TSR)e-Nh+b )*1c*j}0[ș4G$RAuw AHCHX h^ˈiDk45۔%CgLȮs}>i%׸XS_xdGTz{tvTɵ_Rf~rO/¥N[OҎhf\OFk&!V DpV 1.'KxnxɍѬ $a{;.&r1t1-owٻj>`7-'C^xEwt̓|ۮhRJ>k><:|Yo( C-p}xopspN:$યx7Q07p_mT?}`-1tx 5&"Q f0"doSSl.LCk!M(}''=DsjIT FQQ5wh4Ƚ"B!E}Ƚ [1|B.oT޶=?'yqKrHd 3âҌ;$8"` (O x8Pd/{,?f=}`mȟJ ӎS LEtHesXAz&K B_۱+m *i`Yb"' SԵḠԲH2NeZlNhCVaE`"pE5:,)(` Q  [wD B^y,(gܕϩ\U$7h$ZxU52|~6]??!{()X>4Ǡ3XW}G֓8>oE=tqsңotba%MlK$^ cKN7kt|vm{l&\fquwo[  4P& -,I֣ t-,ȽH)'s,gĤr?Ik7<"V)t.D{Mvǔ)B? 0۳;y[~67uY>]!Ktם\ )z*F}GvNl#"d$"Q vIʰǟVmN}}:/{oT7/\qKE])$lWD?{WFOa"Y| p3 ngv1(%(#9/+$[X~SKnm MS쪧b 5I+[ʡ0CQY5۾=2$/hχ"eJUdfk+dHAѡrZfR,)cSRSYl4*1<(;A޽m=pkcF(k1qb X{,S'Ӳ܏[bQRxBAo @0U$ ULFEWtMEY<7q.; ezHJʭ {)=kveYbʄZdc(@1dK{TMٛ  7^N:qr_^D/ ܡ|,I_4wmkB*&W }|ե~S/'7pfu GrZx-X ڵT5i5>1.bt\sx>bl`tWV}#Q=';z :[&͟nU9sP}ĐERh%Š@1B5-nw_<ί:xcD{$M>P$T3_B3K&Uos.EǑC6 aOvy]؎L둘r;롼Xiep.\o=,k&40 ,jh=S5\$S.v?T4 }qLW.PQKWG̘Uaa4N**"*ۻO6W3a>miJ(T'S$}* 8Hvΰ"\c-zRmCOI>yt2k5J`u4TF̹*aao#Xh{Bp煯>4fA|q Ql:|'NO>MϾvĦTg_LZ2JƵ -W*J.\TUk/bkМ=JV 6u6攝b*vҳlB1TLؽs;bǓbCAm;`o$1 >*y+1S)CAiєbV&++iʹ\xX:JcGf`؋N8ƚ\d(ďUixbeb87snΩ_Ej8PDt=#qDěF%jUDH4Y"#FFN[@E NTRw0Z*]Jaêy^L~\]#1ŇX:jutaxu|2_: E&wEʳ E!˳U}?w}1&)f.n)I$# =5 K|֞+`ko&ȭB=acɳB!~g>ͿEpe4\'1{8u^.Me q:"\ױ֘$K"Q5"K"*k?,m-ҎDv7Dpyfpy89IBeWRQRSҮBdѠF)FR11u3U۔q );0pZQE.V PR=O̹}\}͎5NYlJ|- ?UvQ^Y:n~>LgQ_W|=ղ^$,@.?ֲ.nP=pc-Yj,&@Q5a4-{dMA RbvĎg,523ξ=PRhl"jE‚T_yRP)&&&ETjZPN@:d>(rRTǨ*Eo)D z+`LGvCgzpbW|t91w&_pIASUVG5*o,UvI .{dW^sdq2y+#^ۇݐ~4ܾg|^tᷣHkBL&34  cM+5B|W2fj2fPi9*Ӷ$yX_]&#|$ : uo6{7swV{7+ _c׫g2 ^w|^>.\Z'* =JpЮfp0WlPYYWOdXy4N.o,Xfe|vĮ:'_H=7?_~lqj~/3V ʹkdN9|>Ht܂Hɇb'499.ʸ'Ӳ'Xo- ј *iXvO ۜGTGc-lů;<Uw[GM1a'd)o|k5q0ib}.^h U ur-LZ\1n5k4i-zrXH *LADAT-HKurDhQJNfC6y,K2RM@m)AVSi rycˏZ{:vog4דk}gG~)Njq>Y>j_+_}?~l7tVow_ooIOL 21@qQF:J 8kzw=-{@~X3kX\ckAaʡհaFi Z \5s>jwvWk?f0f;r0]nRP)•`0p0pE<jֺU \9Pz-izu**S۰䃝G-wE4Yk>LLx޿,y aya)er"P*yEE e3䣭-mۘǩmzN_?XNMYMK۴:bayXy1;E`ZH)%%jN)0h_ T5f%n7>l]Ék; rxk hߜ$i5}3==$fdRaA C2HQ Y720&n#˃}dyb64Ai =tFh}(uJBBTYku%k* )i<,iscnlEKDt $@yW P"hJg.䳌&sFx;+93;6fQ0J?&awwnϛU>&eoC_ 0֗/>|ZVX &X)dAŠ# skn I%$r'ƃa 3TԐ#'AC89l'H[d>-"UNC n;=tKJ?,~U-GDn;vЈdZ=qwS" 83Q"c. G8g0ŭVv&EZ(^Z͗q0=g<_a*D RT]^(#d%Ij `y<{*rs y&&hpyf>$d#S#cR/V0 w6poK^4oԢJWhNs'o,w$ rP %l 5˼lMM i 4߾MXtGv$OIpI:e}OJF y]׎0#Ns,\b,uc7rqb sD7vwp9SaǜIWp=|vxbp:t'ԍ/뜹=hLα8 p۔7aN?_@I''l2O#0O=B^+. Ed)vmř//~!Z@md:Y~d~Ř2(U01ߨP uJZ8s|T&5.fCAW$x7O~c7ƽovfO_c6P:zG V6;~Ґ7LyFrgW{j!ņL4S[J</'~b)֦S2: \pRAq"+ f.s>҃%oTFة^ah \h'p$_kVpB(梔qS+-طILqv}vM2i -ϳ -:@ۢvYq@e77eqj3 .}h2II)#`#n#cm|۟~!0fvMKl:sd~ ł\,؇ASv/ {K@C*'\e[0/쉭F۱7)=:6&+.&~MGABҾ&dC'MwŶ_innLwv|OhQ^*?G:vz6$?(ٺ w{Χ3;CsHn'}g$SL;Yk#E wk_ܚ;tB0m䮌+v;Mwk.(o-l缝 d =!.tr'xMO4WCc\u/}:}avJo~ջY',]35joؖrhJ;ME+ zfa4;ooqszAqYq[g]U^'gO'{ӅȣpV'7 m:su#jqB;mN>1A˸0F "\0N2%})j%]!- 3ӻJ,s˲ .30@_Vwah!e5SE2c#+3Yk|RAa4g}Nu!Q?Nzأ0G$)XTLE"xȹ25# $dRqUs}f-FK0BXgA+~5vbC&LmMr 6rh}dK(N#Î6(o|fv3Ǿ=bN(Y/*Au2UȨ ʨ.pG2ϖg0NGHUà4JuWAޔ#!!PQZ ,ǘcib伄vidrRpc1_gK+0HSN2DVK̀4"k.% Vx;YW϶J!rgХCT@RJ:ArĘ]S*# ) <^RKv@T |R3D$rNil қ E $޹"Te.?i{>g†3aP,-)XLrȴ)|TLQ wr0U YWfU^{ޏY#́H<12V%U1%5[mV0`H>h[fƥŹ(_59Vde['?@= % & SYS|Zqzf+Fϳb~0k0N9&!jk-W3W.4d Y}qioJ$J* a6y`o +S^J)M盤1EZ jDT2K)Xy0c1K"ce-6*?<1>{1{Z>KBK8GBt 1N(Ed0D\fg"h61 \xtlں_EAmPh;wvsҨ,t4mbVz4m w.}}7>N{`W;ℷ/{u5\vl ,{?]lLHtk}"Z;]%͛mۈ4PWm$oݫ(O~MGNg'1LqF:~fƮ6q~ߞƭUvƥSBɤ gIilY#Ɲ_8^}Wt>h` }P?%Z:6 dcug;ifӵ՗V6*ଟяGͷkQKOW7:n.w7bUf:},ܗ'jD6J.4E{V^g.W~ǿz!'.Yc+b]Sh ( =+qSrip J*3dϣP .Y-hKu2әldu9Jk3j[_%/'i=Ӥ$XTZ:& xr2=jbUEMynH2`I][L^ Ǥ)%, :kʊZ+9yiy֯$'֗gRXzz_eUiGr8YTH$øcGdkH2DZ0qfh*̀׵ْ s"b\Ii3}.gmrV>m2*la58ʶP?ׅ-7^乸]j?퀱},_pT٢%g(IJWKP0ɿGDVg䮮ŖNfEP=!)ٔƢ))ZAe#/YAR`2к;-vMㆋy6Vt6NyB!$zg{׈<}vZ* `o&nì<c!&"3@«X'U&%)@ V{D@Q:&DJ'.1JTRp6A ,M<e`Ha=s 0" yJb9`/J5?LV&BuiNj-ZlI8~^ف Z2skvovd[ΧGMԸͳ2ɇ6* 030N3VVNCrbTTIv\Q%ޡJEt4SiF҇VdJ4TT̹tQ$30+Ҕ\˚| QHG|N Ss.P4FɣYkc!#6\5bkpֈCXâ o.5'M^Zy[.ڻ 緢a= 1w$16 Z ֌'\mFtB,*πd#2΃bЙ[ҙ:vNgL I" Jə J;R@j3n6:' );n6KֺREHfRa.$)DnJVlmyXg.xK=M]NXǼ֞G.;44iNKZKKy UېRRK3Ug5sv.}B.yfVsdGqZ1){&SUzFyd!!+Us)gC)gB`!s4Ǣ$֘+ &"sE䙣I83>JC t6J)REHItgaeHBz_kYȶ$&Gv1!A+ҋӓ֐ISbu{R|Lvr1q/՛iٕ1(SA+6{Pq,p10zi%2fjLyѷQU^&ZSwMT6Q,8e]wc:nlpU%mzKDrb CJ,T@dC)Rlwth DVY'X$PG]DDCMyN6p\ep!Og*5)<QFOړMg\B/)gk9+I4[(E-z3/Ec NX@uFB/Қ}q2'O+0>8l!iK|j!}c$3JT)5)u!!%Rl& H%2':VXk\c#3ޡV)v *!\}Y^sC  oGC3I v' o澽z˛_?oE-a<,FeW"$?/|硵bh̞] iѼkI)BVer$ oY ^[vp<$|p MB5z.ċ!]3tqugd?cn] e]xo ]/\~WG|:'c(s9/9/]UlXFIMMtG4G_i5^AiB'r{zKaZ;C$.2w6W*1漵 # z4 X2sʶ$_*ے|lKŲ-TBT ۄd&!jiH"^)@ " oՉ|;C7镜] |#kݍnF6KGL%0Sjح'iwqsJ+)1]Г.ӡt<*"0pY3]+⚺yt:n[̨&wRrs})^ZU2IL1K,3SOb.՞od):eu)XtBiQ8sJ4\d.YW^r@Go6𨃊4["+X睲kld)dmiY -Ol:AϼVR:\ߛ //4@Dv* ȓ%H&V&K7հaͥ`$W>1Q|r4%}!+U'A(RpBN+kG*Lp.7NLskk_Iŕ{fp1Fc>דmT).xd€Ĭ2A(8OP҅ ]ͯЅ" ,j' iE.d)89keHӅgT!O 4-ӥu {؏ϳ〤<( N(p<'Z!甊˂ӤB$G49 :\lNg'vd)0}J"[ dO&")}ۧ0ۂ SgȌx<+yNVz;߰()v`YiYGZKyF;wZ Pj݊W!~]O+X]`7tU*hNWƁ$7]u$\w+oGmPfkjQ*\vGcFg~h2oM4I>&H1\~w?ݤьM`葪l+Z]R끦kǖ/g'g /GS9 <҃{#O?f ٟLOA[NtFW͉sa}+􆮔6/tUλ b=+0dUlyo .BW]GBFn]oµ/tEh%UE)}+mGtU["/tU*u(q]#]ѕ[7tUBoUEkd骢z=+kUz ~]Uy(@WHWjgDW㻪p7*Z@r-@W}+|d!woȃt\#;P2t]=\+m{CW5}+B Bv*ʮ%EJrg֪uvaXc4FZT-Stߙ ڭ8 ޽-kAkZ'hvHM*zlTlER])A֮:b?p07A8}}7OM?g+1q;yz󸘍 "+Za.+J%bT-`yaj ]U:ӵʵ1;r-;pb~#ʈEkBWVλB+J4wٻ8+D&F0lᅽlƚE<-b(Q&w[(NK)2l2@bV։oeK6PC)u%:RBW֯ *J9>U$R]Af3cWWOW]!]%l * +`BW~ӕL?]G6}81a'5,o<2CR$ffe3*MO~M+VJ&Z;]%=+&jcdQrNINuD.X$1 4a34p M+ZNӊR@HӜ=ճsfsk<+cj`/=v5Қ=51Πx6=6lغR/}@ҕp2od;cW 7mf hW?v(=UD6u[2sKѕ}飅Rtt,oK3v6hӕuto }g̃7WъY=]e]a*=ru]:14n݇ z5?\K &Yo72[.>_7^~}@yK]@\_4vD(ʯ?~y}ҙ=Kv _~\B@yߠprwO0hwXvkog7vzvq-xsۯp 惿0_t[|}ܗqpwF;>)Qvwo%dջw܋W>.,WTG~" 28 E :p13 - z#A=Cۢ>/Ҿ|Տ[E};1ǀ/՛>38bֲSeO \t*6P ~)T_1~ ">me+$k~c uqy|_>u9TI4BKQ$J(֑Ol5|I\rq5-!T!j g/ƵY6rCL >Ov*<[ғze?-4Hܰ&u XġϜ)$뭢١BxK GD+M,R dKhDub%dGԠ"jK-R٬>#sS`q_b$؁zokufWFcWq`6p  MAW:JKYb$8(WZD2 [4LlAgd2 /vYEo{jPB]zB߁CnTzȸCæ|#(P- BH0( "*( 7sTc |k)CNEE; }!!.A US]wRk':ӽ-HK@4`=n `ɦAB5+AtxobMEuo %RI9n,\ BAܢ5대 2 }֑: q ;w RET܋. RDS-d=td0Wz, w u NJDa]$ Y,k9$@HN}aNr|p&ػ &O ħ ^-ڻZ,KLSǘAQե.Cw"'$3[<U&^8kg$ ީaC;& fBOB{T^#'e*d6"t56b U&1:Ādd/(XX肸Г kɐ$Ra2ӚC^ ®wf4 hY3HT_ )[nmՊG⭐!@@8JK W'*YȩՏϦß=!V3$\FΡG o(ohV1 /].:Cc$tԌt ?$ U @M)\C;&ӣ-МV|ghc+ڊ[Vj:zPi@HuΤ ev8k|AjC,!?mB5,!/;G?j։7߼AhŢVqdAiBҢ58m>40z:/dvčtX48Z Fp8.C)Z1!o*"P@.I`â1[.ՂpҘ":b9RЬCV\GR!6OS,]:)Z-, ;ﻺ NgW,"εPRAϽ`j?,Kt!O絸f)nmX1jkIurx GxN￴[{z8ބ>c>9>ڭar d'3M9 t$ao{D$AV 3@9"&0|T"rGR H~F$^ zGC$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D t$7J!@l30U$Z@H ' D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@B J7E+N:GHaI$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@gL15=$b8$+`H Vޓ@h=Β29""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ":諩2ћMTZvz}ޝ\P,w9-VGqv3]>Ð%4X \Bs KFζ[.}UC<7 B7j+lב\UkUw\Vv$W#W8y@w"_;f mtg^MyM^~ ~wUB b7利K%.+_wl:S~^G5 _2lRH+Xm}|1Fw2rOڂ~g57*hc|ͦxoMӎq\\l>W9oX17Ngd_Oʛˀ{r}Ko|,ZmE+BFzSiVȥ4Nj^uh1`+ Wk{0\43 %8 Vv8oF\ j֪5V+w8?+%p3 B FErU]\in^%}z1v~67MO1>*w;ߌvώ1[Jj0)vվ/=80<b7vnϪ1?R~\M]2 ^p;wjq!~#ݓ2c?#컿NľSr_c6Ցk=_xŅS..m*D,(]%"̡ ǂr]^(ԯO1;=9]zY=İ_5v?m9TsnM].n~f:ŀ}h;selr!Wn P?/idqq=߮ma-wl=0Os[-Vvאָ}}.YzI޷mO7Q_)VU^ejU7EyNudR3Տ9\fGm!n:Z`FۭAbz%.G?a #potF wvآ^;ݫ|׻hչ/ab܊%lK:]l_Fm%K Nf)Ā^T4sjje@+z:S u>ֆ Xb߁fՖ;&t, DH's Fsֲv%CGCgx([UvUgp\34zԈUQY fY28 $nTI:+4e:D%W`bIB|pR d :e8=$$ǒ8̜'Xmմ7VW%`O~p_>ܤiX_ߢ&{._m/7cSk V<`-SN[6K~^uW9 B;ydQ[!u ˘")dmDMXݱtfΡpSϒYN"rs(A*VD@v" +,{!k w)irq9棷yގ/ 04/?$:(cÿ"xc"Y%}+aUǀR3`S+ Ɖx¿n¿~l0oj 6_>Aѱm<*KDЙ)kT'Mae6b><< v`O}Zȱ!;y7ڱF?Sf}6T6P6hɉ&NdMlA_/+hK7r2UWLCIFynx td1Z鋉6O3[3YP]s֣{위LR [JPhD8:E)Qyp!yXR\,ZJAtZ~yp;:3xK!G?{`Ir>nyK]doe;w(_vd c:=RMp3k(Q?ZQ|ށ+QW̸ ObqLۮx q͵6 t{$;yH_Tqe^;{wOD S9hgN̢u.׫87oc2VQ漇Gߏ47UנfW%0VWݾ޺/1́°ŲJl@)/M *xf *I^Dj[ is2)17SBY 7C]6-Jm>RLv* `d:$M.,A9iy\s wL%1zx >v=T֙9 4@ pwxãugzH>^=a]Wkx2[G (vOyM] d>NyG$:k]1Uh&ъsFYv׬Eq%\$FxfU"+=<ߚ0YQ"}Bi.s>Btu#PVb/_ȑ<5\qRݙ4=  +>\ILژ֗(Bd8R3\̧Hs4TgM^μ[g0O0/Od5ms?{W۸ ffC@0Gvl6`IN[ӲvE"Ӣ$%[nWWѴn.+H]{2 7&ilI9C \v}|/~؇PRS'H] sH*t(IE̳1(?9ESC޷*UAU#"ĸH}]՞"$^c!T$ܧʾUں6w_ yRSO/8St-Z>AB,u9S xIBVSt3'*kfZӸ?5?_,* :;\?GnbF'xZv7:[ÑGY^ HM5}9 *',W+"=U#}Y41w_%}Q_qeT4w~EsMy7L#tm \60]z7ulTzt&[' ̒ Ҟ.Ɏj-i;5~RL5U#&I@{h5`NNmA@886 32Bn%Aa8Fm[@\`BH}URVaB˥;cƌL $ZFL&ZܦmZgc'C5:#1>WT*j}0[șcG$!)}`5߂!yo] TVF #&Ʒ0:VMȢ T20 h<z6Ā'6X2x B,K*kbCmjb|>?!*\xŨX]vޘTUASA〵r7S]X;I} iO~µ6ޏ[s/hP5A@ec0fOpD,-䪝gAM\r!򼛮*p9d #ʋsZ(/]]&zRZCA!HPͭ;7nȵfKo$}.ן]/{w.:_*β^&hCZX 0 Xٷa)MVEHUO1Keɸ%O{ìϼ'󆻑߷T` SK-sdp O7`.^'bk$փDl PԱOR U7 Œ9K˭.#Ă2~99m֌]6x,Xn IU_wp) O=cF@-h>sbaƧ'zb*7yy)l7Zކb8NVaTq&y6 9+ջak[O|ޏ&[w=-iNO1i!&@RY4;L˕y(]ȹH/9hrF} ߻zЫs\`nQ/Tٗ´k#N- QI)%Qjc%LcݻOʽ[]o^o C!fH ^ Οc.*5X@Uw8zb x{yK$܇prr^ג]exll{O<-& G=0fXi_oB3+55dҸ  >!4|hE,Q 3 WA{HT k!#RQLP'Ԋ?O:U]2õ1$Ή@ 1ZsZ QZRmPG S\-&x#32xN3AI֌٭*JrIơPdօׅGw4ǘ%ϸ9'HP ᠚~XM4 ]`ڃ# G`='>Dʫ0`x6N$KE퐂(.Dʬ[c/'sIǩhmkރ!^H8ĨF ^ x`s43Pl,F)aP I3`UF.qVq BN@ hG]r2'{ Ҙ0;az$x21}.8T#Q׈$^Ib$1YcsAh !K9 HtU#j2A(QHLd$&=cu  9 H\JY#fcgF-j#yJlRr^(^T^6 0]tI,0-5M1@S)`p Q64M1lq"d}xkM,RۊW~L`+ ~3=K }= =>uL"ub%OV?hsf™#Q.d^D$JQ0=[[_8s05E4 #mE4BTD1r+vRu[h^HXG"\JlB Oc41ǬQFYJybB-)^j=p;JΞEsm!̝'N#_s]s5-Mbҩ㳬P4w`t$Ɍy8xaIT FQQ5wh4Ƚgu!h搢vg}4bxÁ TfgPVT;"E RrHd 3âҌ' J%̈``"Ay ssO@^Qhzd<ue#Th wԀLHI1#9̭b*g4AX&&Z_=\//eHQA7HL(L(-r @Q1SMuL)ZOu#:`pXF+D$jpZ2 TQ/C8Toiryf!au 03 wS:Ur FU݇/* cY F)'VuojDԂv:nRUQ/shkUɟ0~ꗅ}x2ӛf"UCɆ%5nȜt'<~1~0_ll;J,@7fNQhF`蛤䲸呔{ b|_gB ]lqEITƅ_0Oy? Ӻ҅fG|>:6f( -~5H7i~<i0(b"!Da^Z#z9EgWRLݙpcgY=7 kC(XҮ>.*lfWȁi-X\?wcaZ'y5W-K6!$/jŻiҋq~F"ZlW8ʶ+\pjWl߃qjݎ&T\W6=x  x,+\Z;D`j>@oTmCW]Jxf^fSJ՟Y}o^/._pmq\*ЕQ+4ʁcn޽}RяG}AUW׿}x"Ǔf;p_ j}nߏ?Gߚs _0܁GuAD?.;~FR~l>GӃWp.6}@ ;5_?.\aFYN(oe]Te+V!6]%wȣ@tjY4-ňieZj UMu:MNkg)(,PK1â0t.fxft+\\T D-q%*mXp5C\ըvFX.hgDR(WCaYj#4%팰`R\\J%Wт)+W"8X.*WVq%*͂9++L9ND.sLNԺq%*_p5C\p%+k Z19Qp՛ٰ; ( )8hwqj8>3\2 mz-+|9pJRp%jQ+Q9p + ˖]B*}>h܃FtAdLZ wLJsĴ޾Gnn\l"r쇽g cKd"bQKn-iufwsm&@\l6d|hr Y͹\{{koFnAӭШ JHmi{|A{TŘ?KPnDC팳Jk*W"P):;N]p5G\[h  6+K\ZrǕ jDJbvX-@;2q ^W<TXDpD~JT斠eիʳPXrܕ5,]%hr97K\֖v(]D.Rp%jm.z3 y&q5Nqrw`ZP;J2N,A .ڶi,W"8bpr-@)8Ne f+ l@>e7N\p6J o嘠%xX?x} `ҋ~[>^;9^s-#W䕘~{ReɩM]YAuUIQ^.Vs';w޽txvi}~q/~8gUle_󝵾meAjGUu%MKՈ氃\'9Wi{"Xi~d?G}eU'&S]K.>et+.tG  R nog@O+rfjP[˕'+8W/fз_OT ;nn f/juBMmg75uu^^;H3DOm FwS::w8i}ν+s5uAoOc 'LY;?\x&rnu!zm?XTɫg;w7|R{_n:X~qo0x{zXPbN[z9dPCfplgN?12 ̈}]ۧu<{=2!X6g,qS~hl OW )yS}o4\'-eKvfdlFHF($T#㶜Oٶ>ŊS .Ns[/EWvҽ]YBA+7]W~)X43{}IY1XQ3KXQ3Q?j*Q o_CnxѨ/Lx%3m omt_ jbi Zx7unck FSAfۂ0ubYm 3<3\{_ JwFD[p5G\֔Dbp%rC1;#ցWrph|pe h_ 6p i*\\*]ZdPTzZp5C\9#  ~+SJb6rEmND%W3ĕ7ֺpł.W".jWһW3U0MIbNbpr0*Rf۰ݾBpw (;1NKp5NżpF-ڶ-,W"xFJ+QU.!~`V`V?ڗԕƒFM ծo>yTH 離Kt^5E';`?(/J0@I#懺>Qg42AhMFmsBŤ\o؄ %P Db}CSu6(&h ߴبET,VyM)Fht;g@ "B1V]:_UgVaV]TjXZus\B,\ JWX%WsO*iF[_ D/f߆zesǕ\,qEVmDpbpr*f2(juAQW־JRpłQJb15NeXqx.hJZj\rYb+Q ޕq$B >oXŋ6j1E$%G _ I eREǔz_UWw铺Օ㪵h 4 L)-dM :Ef}:؟NҼ;];W}-U]_o&{UϱRYf Yʘ&>wRa p3O&ޟ^~|5XJ2y͎?JN{AJdnjh=C3t毰:QՌ[FxeKH\8o :(~4gxĦzѫ~S䓔ʋ^y`]ي'HMB;ٚN \j`:,aĠV.y.NX5o_&z0!TLLll 3Sш excu]>o[Z؏X_q):>>^Թlj^>P疏W+\+&a}$zդf5ϛ><;}ߍe 5Eb(_בr;\OBOV]v*¶Rsٰ d2sYcγBE2`,`GkHFJE)D/Iw7%V}H{IVU$ 0Nhk9 r!XdP8pAJ&a٩ jILLNa2.^chƦ6]8[1`A <4ot [OoYz5=/zYfr%!u;wQ^d+2)<obuJYK>%yGWpoǰlWdtM/W|,Q^G\ }NZutན,N@/0#SLo:jN+^E',MjNX W|!yQ$E2bq+5C16L %1PgmYT:8%ڔ eM&'ؿlSC 4%`^8pU% E# ܰ)Kr Yun8a=x, wq'Qh'>Dn%pr(;eHR l'FZz!# $n(YYދ ޿p|3^VXzl4;.&ML4|Y_Ei:7TWTS8My\\N˥mJV^7͜rG\OOcqizu&aȟz5::P:XIW-n7Kpw5@-,&7M_8Cݽ?e ߖO!Wz͋b1s&Fb`Dy}W;z':TI? 0!z88zbč|rgkq^d.yiQqG^YH1q6egoȯ\32n 2rg83 A!2y:NJtC/rZOEKx,1~YWذ_y m ch^8}EK.{cZI@-Poϛe0_"M#p.@3h:@)1,»+AO2iUEe)sLNyƍ<Nq ݆3։ w6Q|^ RЃz/p[uD'PB%tۼĊ{]CߌͿoǓо-[l0\l8B-X] Uv&mSM&{>k.K+CY4g%4r8(ֈr,B @tw{ 'LPOȮޡ )}w(xH>{a><H8L<8,Z%[ 'bv&yM( pzl@{qqysz3"ĸ9Ahvtehk=!gBZƨ]!ޡ͟ RcaTp 畞:eWΛXy RbzBY}'kZZ$1"#Z$:Z?Ben8lvh ^%'lk8:MyJF[AƘec$Nc-" c.8SP\!ϕB.2KeNr< t<9ܫ&`-G?_t 5qY0M)b4(K*rtLK5f.iubrtyQ! >Nw˷ " h# cNZbǂ9ߔ =RU\ FӛcIOLJLA=00yܑGq'\T֌&T71VޟFҠYa1䁧Ic,F2Ep9򲅄VȻՌٜR\Yj#H`ti"&pC 2)5cg׌atagP]h:օ T)*}[\dNRspݟN>h7T0Mޏ]cCsr&XeˏᚗUL@ dB,Αn5t2+z ك$ AHQԦ46D_&ߎ9GMYc95vS<w6'}Bl  ΐ 1z^Hi* 2HIҝm0kc!pMȐ+DK&!FppIJAHY'ɩF б>쌜v'8 x4>}WPh;ֈO>7`E"*! ^(.U,Zp˼A.#dc;ΘG j%r虐Bt`".\oS(rR5bg׈/:^WK:Xgg\@ONzL:Ht /ɃIt'zXagq$pG'PaɷlRȃv*G` lʆjї )T? bʆa; V,ebnsm$HATY0JKdd攷4y\y޶:m}Dy(*v1KA&'l&o[4vV`l#CKܓ-qD4TN9[YrB&K%LJ 8ܑd;tF΁y۳"c1%i8};,~Ɇ̆q^دMc%g%7tѼUy$+"(RV &!JVZ2icY9UAYөs41%e1SY)ЇRMg \+LQ C qg5Zn?Dua.”;uўMIj|Y/9W6}:vr =S'x[g6;wTx4s 1J$SV<&;ܓuDQMRC*\@A{T9<|HF}""Ҋ`d|:"g9!:49@SF4`c9Yd47GZyL|[c@My}&Rsk#.Hx^ S\w*+_x!IRHRJi ]./f9C3p#Et9pNw<p `/-443hB,KE P.!01' L@[^wPSR`:ʀ_0)9*3ȎJv2 ARLH˓0 =}KwlA[ۃAp=WN'y:wTڅHo]-hT*Wb "J+K1iqcy.ϭm|ȃ |tϺO(;DYh%M +INea$cfT B%⑃n7{핟B: ^ƻOۡL#]XOѮ&UrO9e]b.gVb)=AgTi. Rm Jz¬$]gj{ȑ_n;;Y{]p`⛭-y$ٹp%;jI[v5ŮfC 3R ep&,+z O$zI>~@L:%YP83b&1'\w!m ϦޝY8֍zh.>OKqJl1{LT-qNŨc,E,9Ay;Q=i}ؖ&O={(قo~LƔDܝ8/\ U =j,j[)BJ<% SSU%ԣ}UL WJ%ldɔ:1. -&IQjt#BJAȲ־dBVdR: YxoGXc]cُBW,+Ac`3*阞yCS$Pi߃b>Ĵ_N"/&[bKɖH lEJlo0["8&DkH+x}]@,]NuLrxĒQW.k pEy@ODCh1S9DbT4@W9NYi0XR#c1DJUt'#w L枖BTh@ ]s. qdxqus?Aq>uD(d260.ͬ5^^?.zi2$5;kQf4iB`A*ПJJy&ْhYjȌ։|'wGNޒNZ2d! c3@$:@"-¢V%+Nnax;kڧ' ѡv+b^^IӺJF<&MiS4]Dʉ[+2; ]Re m)h Hǘ1eγ$ėHmt lMo]Y2Z-We^i2xk2SZh=K>AYN,=O$KLOƋ? J4nm2I@*W$ƕ 18k$HW +QNtq=}Pw9v5AzzI^Iβ5g؞6ܘsn̹imUvt 6X,{yt5T3Eh X\)c夕f@b@i(7]n}u > !+PNUYPC"+u3y(ʋGd-gF~yVkϼb7eJu kAvy=p6Vct3RZy G{\z:{L S$:?ٕxD!HM1]2Jgtm;'w^~ Q;e ^10]ҝYr1+rE Ҵ#"d4AZ"TBU,cg ~/$R}!]G&>4C Dg7y,ps5 3  .EtI Qx)x`mS_\Kt޻؟5bۻF561cŔC@L 5qt)80&Rdv󴸹a[gsom?]m{ ]DcLƲlƎUN"H/+EZ'&ob* 9k/<Ҳι2-ߴf LD„T\>^z%+N<ÒG<{CN&&1Š$ZyJ't/ߙ8]}bPFZ_94L|Sc&zQYk*6<'+:c{K[e_̖g"p!,q0>As;ܪ4xгB ju:ee,ǘGT'fH'nk"(#=O;eb[& \dJ@)K:c|dz֙8YWf轂J gA)E42,ف1)0HQ#<@*pZc=ɲKp.`fSyD]TM2W0@ΠVX\KDجNhi5y/r0שaVeiBʦ"('tV M)(HHh?I\p`FI,>rBw!sl6,]MoD}l)yπųeW;{{HZi|л9w]߻`o)|[ۥmr隴-cңeMLH4u1}O焜?|ЫkRW0&ľݵ9r-auHWݒb٦m:!}{f>9ݓ&dEE q\E Pje=Qx6=ryE!uǍVaϝn746(u=a=fVKi^x7Ty c|錑Q~ ԪT,2O!e'z[Uo;1ŘqW6+ob 29 2}Q:2>XSjtWf`Uu67r㳰<6s`}@J\1CLs$[g׵2~EZ>-ެ]n6JK/_xWӻew8&PJWG N./SR˶,=N\ O8Iܭ?0IicJ3E-a$h&]Yd.\Q}XCJA Md}ҵm4vx\ݰauXK64n\F,?S 4F?4W>^ޜ'^躾LY&@7闟G˝'e2{#j0CnNl?_=?C\J i[DZ2h]EX``"8Β&դ oIOId9ABKˬȭc uΖRtdD+%:%6%L,sqH>R l2G#HK阈w&NV6  ĘI*iI ńڬ3qNsQz9aHO+#ƖechIiMM3i dj*"]BIyV^˧73$M>\)tEVNdFGAA&k9%hGy4j%"qhPМω{BVy{G Tβ} X2Xw([H EC GٜhhqX/P=!"WW) i\'u,Y>m)h~>d1z=m 58zyrgk-@}7pd^Bm4^?H&c l0Y9ղun&y>w{v]]]Zm+s>?{Ƒ_Gwkga-!SY&$e ߯zHP#rI$q5t^ᴔ;<nžWbKE"sYc /I~=kX~&C:has0+VM]hrNQTcS\u37u 0"o*@7-ѵ:׳RvQ0=u.$y/%ȡb@Vȭ2N*m7jrVU{M%hD@!X&HSE57V"FMsFV;|t^998U`ed.gra~  ʏ&)6%l9ܦdCnm"XsJkf؅uEy'EC<!zqx/_fIh{>^:U'.Nvnwb~l.&XF׊S\u{d7VXwT SaV9q64qIv˜gݸ(PE 6BeAD@'<y,![nu '\}+iv]ȑ˝V0#)6gF\tD#2a ],NRvp²+x01LB 8q-]JW K!qK08PcBWb%=tqăzk*5:"q!XW ]v4JX|Jh>"qVhUBWcW-A_"aqtVTjzQh/ٵYᓔ~|UzgmEP:g;5˭!g.\1$sNp$U<_Nv,&,u 5EP&quAW~'^5'tIָd/ NK=gLOtz\R]S^rPnR`.s%_aퟝ4 2<;XlB_+Kev!|:QX[f֔~ʈ*/tm#^‹_&~q~q-OcJU S\ʯrAk-g@lc?u?| _םXrkS.O &tdq|h5aIXS4抨w \OmgNٝ;VyIл^uh̽2ghrQaDP%]1#vݶo7L#v(ݶĥs_XhzvHÅӂ+" 紒)F/I@[ono7H?3:Zԃ~.[]`Zl"=z%Ι#+dNiL, I~Q50hȣ{aMlqؽ׎ Kx\~<l=oܻ.nYMsWqTeA6}OU 0mD.9c Gf0YQj v8Q(o'T`cyg\k@c,) ň`'lb8a2L=ՠCPJ<8VG`Lh4!3˙eFHL]o+EHD"# *H!UR ahxM{g_zHl&y]ԅ6" yMq>0xƘu lH=P'(T:,' |RT+iE6: bY,5"j,Hf"sdnf~|H}|ȶG{f:ե{yJԊ@$OEFBDjiÜ%%R ;#CfKcl7 ٖ;Xb4B:sŐe3/ 0$RI0$|dhd_&,S׋a+X0Մ_%7ihFAeIJ NM|FV3ۼh$g`:a5hߠ5)T+F%-r$$;HO&S1G5rneZGEdQ0vbxĦFҳ=:sw'iOZq(C QޗGj~mSا cgLd(ݼ}ۗӥĮFOKѳ蠳oxΗFeϣZ=ؿ2jQ+05d6xO48jXGBKăwN  D3pm/A'Um?vX(zXl6a[$"*(`$awL'#O_9'X0)*|ZTW/K_2RR n%SE/hijI_\z:a6L'(ĢoRYc1uӫo_O_]s0\;p;s=zo`tJ~/qUDfX/#iq(Gٛ~U)?X;5oLXb/ejڊwj~ w:]Oukdӏkn)s6Mǰ቟kH]wWf7P o0kX`K{WKYi/EO7Whٜhٚ7 1>Dss~$m\q4"a"(< W?[ cx13Np!\6oxld;l&`p=hrݽpm5p(Mk- Tzmc|7 &5/D )2{e>mxQ ^Xx4oĉ}Ƈjjl͊4d_<mqAi}i P%%©Ӟ0p N&Hv=W|K8_^iOY=d=s.aȬ:#y(̤R^τuRTLjH9G% t"?V@Zr ob=څ/tk;R*qPhЈ#607b }OgvBw{B7ꛗa+m]`7,%_ 3(`Vdr]YoAɨ-q5d(#jcK Y.rYgli ~!r^˃zb<_PoT4.w7ɼ#xAo2WGBYػe7Ѻ 14-QM?\e9דwz.+x|VߓGP x&FiCr̵q8QC%Z Z% NV2pqlRKB5kcO Xhh>:2(:T_Z30k1hFs勞!}XFA W+* 5>)eL αcJU*he0P@jM/ހִI*\6{}SNҶBݦRTRpSe곤WEӧ73f(bD9?E{51eF6I+]Om>-;lΡ(#[ZSze@q`YV\z%:n0>Bpɍ7vIQjR"L3BD4:/Gf-u5%8Rjȍa6nW?}8-4[X/R5k֘KҬ_ڽ=5ϦP/s *鐎q`3 `Wޛ^%7 S3\u37u 0"oV!7-]oIrW;R?_Ipr$E$0ZiJ&e{>Lʢ$Mid2՜W5U_m oW gIa)A_e_PCil%"}Xʞ!PJR*p rI-6&m2:2Z)58q G6d3)Ȩ^̫S[-1ƚhkBw5{R߲ٜ!}9$cG.'L! &y ktddND8J1{t]1_|KnEqËofAE>msRNpwFs\q^BL IX#YjS5eR}f, "=!]*x:id'9-AEHDE٫*XI?/>TQ5=DPh|F',(Q+(ViE=hY$XtQ*!5J d>, DYH?JI:Cae))Hw{ 2UoXR/GGN3KrEM\5E?T-.g5)mrqu te_|9g(Îf/B[,/5WP0Y}eov4i)]iNR= +|̉zř˸_byOYez-6v}yb\֟o7hò?c|{1+~O^nqRg̩МvOmӫS5?+Lg|7d\2ᴦ,Q-)&t53?YʜU6/l:em7C xySDj^oQ'sWn|>~FB\sk zN^4tmkh$E$l4. zefȋJz.SRڗQ)Y*Uwsqm=;K*[٪8,}J^LU2GK}ޥ ܦT^5a+X7obt24]®́@O7[VY-]'QfZw6ge?&,s j.0R]L1΁[}9Hg:vWʢC3jZVWr4bEeM;ZמCa.ȸLHdƱbY&=0:ٻ)el0vcIyP.d V֝.TphyЛ\{ n|ZUpK0ÏPhP/_oELQA8TR]":uE.`.]eLŒr7»o g~HFDI@>@B,:ſB3K9(]&z{@{m;қt>8ԵhwZt>w#z"el^= l/:3x[emlFفԡpQ2PmOO98O9 O9 OEZd}f<9W+YIpD5hድZH$w1@$:~}>M_7TJ `t,hvP !ɤm ɫZ(-k@ L7 U 8ymlal49:'|j͖sR?ؾ3,Mnai$'Ӝ5痓>/2V2Jtі@(Bzʜ$b*N>N::p(ZF]ɹ(jm(6=|Hٗ([i9tjm-~4$P[hB3£p^xό.>,t/wp:|?X\cPA:++92ꏕF֮_0:/TQΔ$}[.{1*%[SvW8dDE$˹bY 1c[IP`2 a )iR Ai dcA9>kjS:8NۚL"YEׁ3j"qPM=laԯGjv㩋ZwAXN6DiJBJDW( ǃ% AhCaڦL(!#HNEfNv!(KQ2ķ3 ֨ RV1n_16]]N/҇!g" iSw6 cU Q& fD9ZLATWsr!Ȩ)ҔʷO*`g"X,Jxˆ3JU*eT蕻o w ԛwHq>Ey $"ewEJ1|ʃjt߻1̰OH9( R*eagR| 2T콀Iђc=؀tlN0Sw3=t(My(~ڽ|@)*4 ; ] ;+rG(5PJ?Vz\?vz\?Zz\!!),  ':Nu#g7Q)iEIE۔38*"\WAr\2V֠"DijVAbR͖szm *-~0#ݻoִ7_ cc=핟 y6\|"%* (D.eCymD&((5-97JkmF{Xwg4kD:1yLN 2oZ^C:v?HAa$|5@ {^S&aL>XSYoC/WkP. *-@CB`(I3VUٶ6t@ʢkE( mbLR"(>?Y'ti-~۱ZNw>-o~|xTwY~~? j~>FDƘ!foq_,h31,ƣ ;o4vRQ"Z!Gy$3pf "K3Fhpڂ:u d"S0Ҵl+B@%ɘI!IOL11 oeKkʶVy@. ]5>tdۛnB̗U pYEo U_n `>5U9^|7붊VgԹ:rr7RhdS$N*.* YBŽ@8@V5kdk )R*Z1ȱXk2 gQ&:+(d$YtpVM([>cb>f¢ Qu*XA7IV!$b1mRq#j~ZiO^8e}ߛY׮B d?ʘ*tfݮJT8PBcҎ6 #Ouxg1&RcM✫j6dWm&)JXwO3?MҁO-氽Ww[P<9{kHGyٯ|/+1R'EL+f+6Z)'x_ rl-+m+#'j TO~)ณ83/+$ĵ?8L{{dS&?2G2_o 7ι%L<װZ!)Em8bDo7@o4?lUp _ýʿo4v]9"r}9[̿9nߝl~?y'TrOnc#fM)#n)Qz/TTXDO2AڐT(vK7H+ S eRbqN./mg|_zSko.?.'$v%K29|p6^&I Q)Lu 5%|OD>A)ΒcOBQ:0=CBXg6kùV8[y7=;=K+'ׯR ]߮)./v_,eId ``(&(w ZO%Q (if \p; S W (=;O5.7K^&L`WfȸWGf񻣳Ӌ˷6&sm9"~LGz Ái4 4  ߜ1v؏I+_/w|2hN˕mGvtzWtÖ/viMl veaU-v fъח/t ً$$o_ VGi4}$$EWDʦ6"6R9M$$^ 颻n9=$vRl:Ya%gK6)֎-?o {Yh̻pߠqgl1\-ti7&9+F7y|ӦotOl/>xi'ird÷S|{5 >lQT-+[L'GyT }E,ç8˚|Goؼ6K=9? 9y%l%X3uqt[vl!_x6s{ ?Oj ^q;{q 5p].>z{sڗ_߯•TI4|K(qK0%eǶ^/q,5Bm.oy7\_lE(w-Ѕ+͠Yg]ȥ4[۹7L~Q\lrHGV5ZI̼ihCi&H ĥU F"Nh) (hѵSb\O߽p2oYȥ1Xa8Cd4;\G一I$N1Nm R}!401ZOfpO.Վ tm5=QKƇ#B' ,P/{[h2T1ڀ{26dR)C1i=!cY81b5%Lw`}P # !0I.C_k&+-SX,/@TIwƿ!ZIUB·b4gU`zm5#(%5Sh)=@n$*1Gp_c;K)kld-ilHn`0,yUbZ9i!GCf4*ܜYB Yd|nPt=KE ٞ;A0}w Ӆy jx)'Rb@# h,*u0:4sܗGtVjD0QRzs`yUo+a 00! -#1yR[ʺ*Lƅ̆) nAhJ}.sCFBH.N`\@QZ7g=akF ,Vd@·6P[ Jr $lk] 7.a` X56ܑJݑLl|S}L(pm xPZ3X(F"J}7ipӴLob7"4qvQL% eSJ1PfP57kiN2-[fkܾukC7a:?%A`t Yࠤ`gN42TYOy0u+@e4ȃyC{K2Y1`(q Ls'E AVY1!=P @dͯZ@!O@n|A)Erj{URR}:gu`e0A\K_#\̑Ӵ&So#f6Ue 3'i<ѣyԄ _jy@V24ETLԴ&bA>iWжYr0F\/1 JWle#-(w2xp3O_ .dA:Xw3umOU;g@vV@*Hi`m1v?nH:!yQ'k7مX*ő`̾ ^{@#EzvI!@@/V9І(T&䡻%_!80vn@׀R"42;EzXKV5Ⱥ-X-&*gta@RE4~-3BL2xkEHUAde!S\(- AjQk}."lByظ* 0@>w{%bl+'jj,Ձ2Zu\!]gѝ}rlT%@Y͓xp6RiM譙+ *"\^vkIPVN3h4LJHC`@Von6/{)Ǜt}L{~|zү1Oec5[9x,gO@3GφkOi`]˿0ljq`HB\|֚c";j8τ5rGCo m2FnP^p MҌ5.\":U0ݹR@S  ؚ Y E+`Ãz;uP}V)Dc_P|r V"IZR$݂u(;:Uyca5 QrO(Bc$R8)FR<&[z@U h5H?x,ImL 1@csrp݂s3Yfj ZN&_#f @Y,d +HZ=˼^f1Xτ:Tmؘ\< 'o9h W>$AiV`o#0`YAZ #C+υip -j|d$NUlp1:jKe`U].i \TTb`&-MP$y xk5c $Ti`MO^F@I>JwL0P5v[FO^rzzwM1hnJHu#=M٠Ns8ׯ~?uXA}oFO,w\͞G}):O"Ix8J+t0JDғU"`W|;S': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N u:cCry'ćB|87!ZQ].NgUQ': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N u:PM:$'vp@.q@Ic @lbJA@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': NgbB9$' 8f`>}'Ќ2Eu=G': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@ z&sOV~{;wTO?\<P/e\BǸp#q Ѧo\bIFK¸  8ك+ s(p5jFV•fW3\f=uQ+P+C+S. gW3Jq W3 ˿_ׂC+D͓gW}my}>;iˏ%~i;{ vdsӖ_v{?/®dSn6QL8kLl|^ ! [RocǷ_I?X|fU<iouty}<{֮?[6-3 4&L ~Xヤ&&G0=/VmqhFRsk#*3"Jh* JzD3R>EE^6ye9C"iiw+1y 3R&jD9C3p#8*9~Š{83o[GԐ|zFxk'hPܔGϏs\G]oqR%o@VLNb1HP8r zjTS6mDkPzzk^Y6kˋ%7+ZT>M8I3[iEvώdӆn^v!Aqc{5 Ji37W !!U=:z!tO> ɯm,]ٕ†MƗ5@V MI4JxU OAc@SCNxK{ۑLVw-/sZJ8 H{iw}CX*O4i4kO@,9 gȥmaX :¾?]uM@mF" f4ʥ%r{ o{f߻Ul&lvByB70m}.,G>ISo.[άAmuP[p*!KA*ìyEaTz=`ty-aew=,a2,8#?Cw~||*X~)k% b6ר R@c8J^{'=.=.=.J9ʀQXR*^g);/$z6Y 99Z`3Cݠ+ǫM~8I0[Ԩ ]h'$j7/6VyvjOovQ̱bcJL0E-qNgf1ꈁ肓%[y9Z=6:MCsqm-|ۻc^ڨS˾i[δ ܻ,{fzrG+KϸXжQƛXi>IO yƃGo/(:(*%5>3 SM*H$(P45B2%Kt. - L0%3ĺ5zVĈզs?F$ lI7KӁاiS\ύeR]oId=wG1_bzy"U bYM$X.jkmY㴄$"6.kOҚ g>Ǚs8wgNLQ*TlZ*+őS6eV%4`JACt2T<2\Wř5UęզsD\.w ]jZIq^(Yꕺ~j CmFK՗&HmƗ0LU>ZCf>6_d1J1d ¨?LN@i,\u ȞƞȞ4@Lh,1::jd #KL[g(` dktfdMRXWNOr?I/3/fv~yRIs&|%VԮQĬLJ\0Ziɤ eS1U"cVkQkNg̀y@uDtuߵ{Mm5Q[zlW~=MuxP7$*Fy4Fl)]m~OjωLWލNN;̀19{vKdQEaAT:*+i :w^<K憵J !Ӗ$Tp{ )F"r j sIR MHS:^:ƌ .s!m ٙ Rm:eV~L([eJlK'/1XBdhzfyfx`ٗ ~a0̾?eJon- $DO, ,HW@Gt2,)$HWzW@ M[q]Q{J=.PFmmEw=;^1W ֯#rxAx5]Lt5-]Llz2Kg.&ﰋ73VGDˉeM('uLcdR\7Lh @RA;vOw]uԽwk0: 9 ¹2#RdbH 4JO }9h>BXc%Wk#Vg]r*ؤ_9-ǻHyj$ BFmxNVFt`y![8}¬,i{!sq΋RGTc),c9b6#o]b3skri60TJIǢr71&Eה C  H4RjB{_ /r$΂dwEf d.*ghs49Peo'v+ɫ :J{= Ų'uUȴ-'LJ=h"S1EB0^pT:ݯFtGG5嘹'*&2`cWƼ,iI@%cL IU =Uܮ')e"i]aj$r*̷;&cP2`,0$vb)k8]kiH;Êc u;Qȩܡ*szNiQa0}#ҴCŐ>+ Hc^(6:[n*i0i^ҤQC}Q! %Z is!PA< A-nrR)zTʣWVC&xҾҸΠwJZ-9hffB̙ds}oK.%.›)Q~uˎv{5SzmS<65}/[[ 煠au\G?W=Jj4٫&x5N1hz'.jzɎ\[*c5@Zzت7|}}hVXPZ&X)d 5(AGA$, ּm %I$BEH= T>iGd/Jd$S(MlfMpEK٢aoY0KsBNүSBxqDSF͢f:I KØK:&Y6Lq+-v&ERRE -=Mf={EO/yޚa* +>FxAtA(= G_ fLR;d@zszq"B6#i: dGDІ8a>$d>GFAF) H(ì,$Ix]{!ouZcG P}z(m \VG<~R !^x1hwBt썋qt@fZ%(0:cƓ942\?֍{n#F8bན?N>TкxJjXrZi]$a~Niz>"]XxQH`Y3Y&h)AJgL(DMHx; .%Oz9%Wq۷_ի8PdDe5Ѽbj SL% slWP4stu7V͆=}^o7m`eI$hp$N3%4ku?{ G U&@}>xP(=mg!Dt79ą6 F՜hV<^=l4d1LAn$62'똃5IRc@v,I}yHmuXn ŒYx#D3eJ*1Ѻ(ZafrHF!|V(Zg̼}G36B.yYUo8Jxv(.XVAalxrQ90:n^&.Ө"YR$K`b]0efuB/^QTho?I1hX\92ވxh}PtmLv6NP%aqǦ= nT{ۯh5I2iTDm6h,’ܰT0J{ X)D; LNUe;h'>yJtCn8|||#®ޔl\ŪK׮=S͞[at/ R+f ޻'wȕ5H]9vĤ) -v[MmJ\:v nW ﭞwpW0V^wk<zs=7na[n oxΚMEK^[wدU<L:`m1sQHC̘mEڹiNfMWUJpu)O"Rg#:Pqm'FC37s%)z0Kִ@{?.$wls哓G/ʗM?kMp׶3j oL.8Z=XPKgJ-Lʈ[&S .nI}Wi ʵ]UL. r -7'ΎH` "lidNڗUcBkJ/w`g_t{ҷ]F,53O-qm7)ƆL%uIӲ\4VԒ 7: Ndc hJ5`!$-y;ACcPgiG; u:dt 0Y,"P< hTN8b#"Өu#\1$d&%L JXOZHC1sP>s/ ۅI[''YLj5PÅ e ppJws8kW+v~G[4? `o+ Fse?b*lRuk 3WN>#YzRstaP:BQgi$yoLuJQ?yᆱx>iӋq5"j.Ͽn9 xn<"+/6 CɊó`Gָ@- ?٣{Al:L4-j\br$ş,چ@3O5v!kuEt$z]10,ϗ`b`z ? ps\fCnyULj'7DnY^2c|^x$E0Fq:xa]gI:QOpxN+ )G}L308wqiu^M`H\#9Zz0>6ez~%/;.#D"ر &1H=גiNR˔Ї0͝w &3&N0 6JJ4ܼ ]7^ p'@XnYpFEGU9\e ^,Wg$Ȝ\@l饷ef H#S7͆X$ޛI*wJZZeBtR+e`YQĄHJy4$ \ec\egAnq3Rrk)SA xP< ɨ"?NP \8*wc@ e{S/fnGKl剠9Aeo-tye:(8k}} xv(8g a% (?zSQ5S/5!JgiZ-"{P6sKP.!]X)R̻/ʼHK;Ǎ4ᣱ&B'-| ڨ/Vzcb01`u{'<LvZP׎VKsYZ+Kkryi44`n'$ 9oxVYsx ^_f #6]&E0|x=VǑ'qnZsqT@pº$-S@M:u5rvkY='Em@uxhgVulyF7<kWE-Z,{WW/}Mٕk}@;Qբ}dXxuQ׳&Ӽ  WIGpcג:~ٴDߑ9#)w$pL~R*ߟm\%,81P)k(7!QA;͵XڭdTFS1x2I#uֳR49 STTDpTג3r*8cSY: ^{Q=DxI⚘*Lon0}.R-%I E%E\Leet+Y0"#{MtPh9È 1% U9L6Kȹ_b0.;#] J8BGuFcT`S) Ι8MÝNaJD8 xс' &&K`},0q>fI:i:s<3 ƝM%X"Du,;#~YW@0jDrBԥlv]\! (O;#@ }WHmNj+r!pWOG\)Åd{$`IǺBrވ+vQJy]=Eq 쑸ٻnY0mQ h1ke$o wx$COS> x-PoÙk?*q=ς":yUtk+癐R \q iɳ+Rq++WgW$J-**.pU'ϮHJ ~Apw| ָgW{ lKu/}a?iӂ+\}qyVpE+H\ĹUvR^#\ NtL]2s\I3S{[2z/uhJSQFTl2?=g\N#kQd9i65h)6v|ch~=ѝ9 u0MI1&$%YlժA̹2C~CE 'oOą&FiGTBbE:ڹxZeF?۬m2_O>TFjnGbS6Z%hZ&rWuVG:HVGzVeݥ`pjGtV~5AIH\g| b]&+ZƽWP"h(Uʒ@\>hTd.@ 2^8;VjJvQ:OLk~Xp`Jq<[[Ipl0XRbe}?l^|K8hU$y8p "Y/x bc9K57>IC BˋiC / L'A+qqZ2#S#c *uQY#Y>Y)Hh3בx];wN(}jFgw5C!Cۄڋնa԰=ėjxQE]Ѕf:iøW:HV3nTy5W!Zs/#8\ (b s͓-[feoVBg%HrsU%.eoߎ/AE߷R3xgӤ_Udޣoߌ&/>9ZV/~&uS[]2>QL`벧]?w}i2_VlhE٭n6K5z0궇ŮXۃv,.O f=8F0|xj^^\w+,n޽?,^8ޙ Embp3xͰ<\QYZƏ^L>y$h{d+ =?4{M{o*zY krX6qպ%gAuJZ8sCT&5.yͺ:Wf 4nƧR 8<u/+“Ǫ3P* = q@sCs؋_G 's84k8L,k=yOHtr5"+D[:j/uΖRtc 9xCUM R2: \pܧ2D:4솳Z sSgT.1x:ܲ?5sk7b[|jJ>d򯽥[QyiC1k{>KA?ŶyusWZ7mwJ\{7nnzk\#vaJa2moyu>]ZmsyηtȀv-wQ7oJ?u㺾oi㟱m߶O|TdY?,mg¼ #"9JE$l*q9JEEZ#ORQ҉R+T$7k oq3Ƌ|q5 \ 6mT/T2 8lx ZON:,@ ԩ(zOY-asjcH%i  Y"B}fb4&t1XD룂J]UL=Hrcm ۻO!gAcŔQiTrAǬьГf`Lҥ:'uڜ8 {/j_Ov'v_zq~uHI(0e3Vrad#lfBU 2΢zRetZ΢ZxaI)XTLEYuέm5[@"#LHeTU $q֖tdAj!N&LxƼV$IUP`~*jl'tCOɂ5, Z_$_cKAFqѕW>JKA˥]/v|(vSm2js2hIG;Xb*B<[f-heU:in5=p4ӳb AE+,c9K W#OfH'A:2H+0ĝ* -Y /\dEY;\v)YgLH{YYϪz)9W&0QK?@vQJ)XZ&ƤRqH#CFJ-](s5Uӗ2@N'dYY.LDJ"'Bf d.*g&!hz"TwI{e3qCYiϳ'`P,{ZR e iSBlTLQ w*qReoR xe:r,&2`c[Ƽ,iID%cLI­6 + (;m!n&)IAOOEMCOVVe.U % & RYS>AZ8JamIKiHxV :aE%cRNA;͐CCU%`0X 4m4(_@xVM}IF BJzdCdX O˫ fxh"fg<{dGHITM&/)i䠙5V0R֓9/L9Ǻ^y5Rj@,֬uz7Ye&w2W>  `KlFc9Iu2әldu#^I>@ QEũ_OԉS=Ӥ$XTtLKu\Ӊ^XUQE^0o7cIzyyT %w˜c9#~Ҳxğ1?bɡ䉔^ٟW޼}ҡҤRL/x9Eʖ!҂P3-G d|YE"shp1K.z:t새YMpC 2Tmd&vdU653BS uG{Af7 -_ܱg?o`0yПξtĆV)K fD2([K@ArL" 3rWYm(DdBX49E+HLsd%+7Sfh]݈ZlG0dP86jw vƀXpQ^+BrNBE!C)Yd*fm,`p+K)GEV4d kbbhDQudHFuP;ڮ85GdqHa>K_ ,%d$ȇ`Q]qƖ<<^ly$cSdfl"dUTpu]x[t@T1*M ADmQ҉`IhEiD ^ a]meK :Uj9&h)1V4: L-t;'FxJ8.>.q>/l$'%8B.ctQCLdШT$>.n#.K<:$Ώp !nUuͼ~| N)]U?.\;:,HyTA@h)@v"lc )Pʣauz@+nX]Ⱥ42|)ѮǖB="@{0$gGV.G1iqNֵJ]=!d* duװ?G` >64˙T A(=wGڜ^\Β:taR7g̑֡63rf%J@-b{D@-,z?|cͽt0nR1qKg"u~&_O6O?~V&jO,ӿcxɚ'|iDoR{)MN<|yuu7C xDj^oRG3٫I7>?#.)*t ]/G)OrGt BЕ(1 A.1. ipr9oYenJ(AikҾ_>?0ak =_b;Q{_->#O?ܼօs+UX}?O-L23ә?CMS2sy de`StP؞k&aS|cE[:S9d?e#3i02wl˛bZU&& ܌}&LAgOWʢC3jZVWr4bNeMdv̀f{,¾03.E(qXdQN]26ECT9Lu7h/l╈0?%|w.zٷt8/RK9ҋZy_eʲB}nJqO>Γ C_xY@w?'_ ߊ L & 5\/OFvGhھn(-/_Lf$#gQv\βm_;&n?}:8 F6u:u6`鈷&/ow$o[έFmlG:8=d姶hkqŃZlj;ӫp2 LzG~㜢_^W5vF-Kl/ƖC>s\쿽g]fϞ߆ VǾ.*Տx6]lxb'swC>ׯ*,0̺YMU.M2=3}!fZk@7*Ni]G.B:6fv#+g+gMqX`$JZQzo"mD&((5- y;q9z'ߕOz_M |{}0UBo۞߼1Vo} ֱMpR!]ơq `På 0l5pbZÉ=j '5Ğ~zIZrY8 Vi#oTF1 $͘RXUmtJv{snV"+Auqds{I^n~qyK]Xz5=|!i]kFܪKs6fbXχuQ`!X(K*J ; hPL$8ƔjrG~/ko8!H5hk)\>ԧ_&he`^ `pM8:[{`^GL`̍tĖHmRct~ġ!䱉 xS딟vvvsI'Q90'Z@CQV25Y 2I &1)y";镬ŽU:q}C;bd;~ݢ9+\;Zfw[eؓ"sڃt/ =Ͼc@f:<$ GE-r&y>hz:=\.WT~ۖ|}~`Rbve'||iCGeڦ$VMP%acj=LmLV푎 }Y1d^WB`)"hdtvY94ɩ!xp/oMBBj=l,H,hm6Rl. I~ΦD&1τDQ'VId텰6"9}b͑YMpAt;K폚 :& F /l N"(#gTβg<=lx6JҤ1vI,̸23;' :Ԋq"o_7MrtGtcl]Vr'UL^\ +~_I ghWߓs/>ױ_`>;׾)yF _hWBD "5AҀ(bq% 3 }ˊ}mYLEɐ (xctelBL%H9M+_TW`{ lK|__v{W-x׺;ZזrW}O`hIDmbl\oJ@4Q1j݊ M>bLb~J$ B(e̛rIj8{Z ]ǣѴlʜ,ga8EK-=|fi%78%<,5PDjK"4BK"H']0I/PU'燫:xpf•W*@xۚ:xN}WuZ;6$_.V]ad7x_.`&+DY{߯jHIcIt"t`$DyXC]4h|%׀]1-@bʒ]=#] CoABJv;U.t4Qc0Tf2 2EW ^YA"`]nPFBbʈEW=ԕ1 lO<9r|~UՖRWftYH'5:=Tv'E ~K't7O(8ѕV5pSꝹ}j]]4UMu,hf{\*rX`~6]|1?^0Q+Us&pooD8ifP7w cr'_3VrVi*՛f/cwk,VՈ1.a%KCM]ZWSf+F0jfΌ4ne]Dy'&v]+'O#g16l!#9A0dpRa:d)],paepr@pbtŸhdȔ?r`H@ni}SƢ> tEɮ׊ɮ|bJ_FF+AIN؂+ b֩rآ*x}+ꊀdW]1-b"J>ܧ7*_+ʉj1A!w]1%쪏 4Q*1dwŔLdx>`nҀ 4pQw!*2[3etezRA]1nRtŴsQjUtG]c YJS@C>/aYf*-vh/UjKNN葾)v;jXJ^ lkϡf/+hU48 ǣQ53SS9p8FSa {K &Om8u"nTqD?7ʍ:Ukuv% ֈIzY)I/F{˔%ak# vKrt)-{uŔ]PW.V8]1.]6w)2  ]ȊbtŸFLci]SB,ꡮ<[VW ܕ׊uŔ誏J(IWp N6%D]PW\p9#+g:bwuz>rp_ʙlXWi]gWi]([G g++WtkM銁+JIӚuŔohpQ)v+#EWLIm)]QWƆҕQ<)&X_6*VC.%Ƶf~8p L)' %E֘Y'm.&`I0K10:10{2̔J2dNZu! >.fvgJ_+0)Nߝ89]Q6jҖ쪏V銀A1b\9N6f]%]PW7]S^w+ubL|bwK]E+wEZɮBbJ(꣮) J}1!w]1.j)"Zu"JKWlz8UMnIvLJr[ džЃ ]iʢ]1RtŴ(EW=ԕ1)-!\@G!$*el~dV1NOoe[ezzgw*sN ZPK.1I/vx-X{˔ޕI.1b\PRtŴ]'BVEW=ԕHa(gp3´6{]1% )"`A:)bZsS]GWA#HW:_lFi+e AE銀Qjg\imŢ*$]10i 2nδ}vE誇@+NNvŸRtŴ!)4*j]OdH  &jFi&']=6:B{]"8\|Wb'yJWLUc02PtC]&=E{Շ:kDqڨr1}+Z s@paMS,r@KBdRPV<%ALdbaƵ %fڮɥQ SI$9ZR۝;_♆qI1w]1SEW=ԕxA" qKF}W#SlUtJRc 1bZt(*}UPkIN8ɮ7(+EWLՊ4hzmuyrbtŸZ49uw3%쪏B+r+ bmP]1e&t6 }8P I6%i]HOu]QBfBBcC`JkG1bܮ7M1w]1eEW=ԕ1| ]0(/FWAuŔ]RWZRew=*:hW>@ޭ]3PnL'T (w]1%*Zޒ_51]LSzUtC]$eWrr+'EWL\bXGWqbUK= w]v՞DtfULU,zlCcQj+FWk]1)+ꡮ52c=s9~ƨ0Z Cs@Kpzl(WfK9%)֔ZDВ &0VP2LOaVy-%&Z"$JSI{ [ ]Yg1i' 芧t52By$]pN7jdZuEQ]QWӒtA@.㢑+@.SF+SF%EWw4)5]PWR F8 w¸VLvŴ}vŔ誏{Qz 8/FWb+]QWVF]bF6f$Jbճnz?s ۢ\NFz/>)a.ꤦZ??WM K}>]Ooj}i/ T]`|q 6 ~;9d:N.|~z1UG.n;5C׀Oqhh[Fj?]٘˷K]䄞ŗC??p࿏ώ6|u~:;3=]t_>#*D?^_pՍI1*tԭ5b3S:|y1?22jV|BU|2'/.݋$ڱ <O|5uzhjurmܠkl|9[iXlf?"k).?;>,~ (xo߰ynۧ'gTtG2/nL'{Fpʫ}f48@A-mMzMB6`=Z1RҙD#y8|zA)9;^_翭m3$kMJn=h.*.I-40U)RoͱCUvzU_Ql@ѷWjkM FcˈеDZnoK5ӘHI7nPP=Ŭ#XHvlz4=[~X3 j:)OZGkm!'d4R}Ft[.TZ=|m8Nqv[kdv"]Zw2OUkGlU.t]?S!k&L c1beg0F޻dt U*G\ixeU'gZ߅X{x҅zӅhxy(ޅ#*F>p@"[tV`>+YZCm ?^V'ٴ s1UU!kjɺVBr{cLr|C;h<oT^W{LC덱kKjTȻFx*4<6޻6b%ڈ?l̓,xn=/bGۯ1YHY`ronKn.ڬ&ulk OfvʪvA&۩6VEjώX<*ڛg5 ޚ~V3ff}l{4ܦ`wI Q>MЌ ~a='w#*V0*7VX`̀i*]~≼Qg֎*W*dl (* Z~+Ge+r7;/.ŋ ]tzm9BuMm6~x?BY)vyj=<ܨt˝M^]?,~󺾬7ڃS( -=uГzG&k`lF#pSʒ&`#|o ku=mNrبJLGX|ћxU3ͭW玛.[ul<JkG2Rh)wG+'XԖYEy'fA6u+b:{kn_TKM?7v"eA{(~R@r&7QI[}nWNFH&jJ^ie}CR:{`1&&:4[6?Nom2wmH&E/$wuU˞ Dɲ%9/5lSd RǖI4xO#,alj3WBJBdP`4J墬6Pn? }`I-Nwcߍ9G' P"Pl^U#``ꗯ_t7{l\wǏ_b[Qi0` %C2Ȋ^޿|^Dize&$tyUԜ0c̫iYhQ~ @X.{L@9A# ujn61~gZO-@`K]YRthrQb@-3=wΈ I@Ĭs9iEr”qR鋄ìLDY:Ncy!1҃LBLcƂYqbA|Ar -ť皓 fiEYCWR1:ySAI"fs(h}6߸~UkeՆ3#\Xk96̋~$ƥ* tik%|^΋wx?Z1mn<5Soz5i_;y2cU [B0r+ČJS]L \r@E'S׭n:A<5aOU8E%綗O>kbWvy--x2©rhRսRx$A7`5⃴Lqc'O&_~bg{Koͮ!&gѯ],Wk[t K| ?ř3Bi,W~wY%|U" = QBG~xۙLa`]~FT뫟y 0\Mږ}hZ~ )u2`du`ܛ.6릟|/^*NH녚&_ZP|ő K|ϵyPƬzJ&uy[ZQF72Ƨ-5,MVjq[ F4vj$0n)\M,.}12kNbѴ2^E[\ul~+O fOÄ""> m=v0u?~a >u=%' !ۮ n,.|a1`t y7^VOo`ASwO6\"͊<ˍAXV֠R\nxT0'kԼK\&pj>mѕƵj@!rߠ+N% i[؟Ko#2#' 40ؓ_Jx5|Upvlڵz}l`wU.f\~[tru6*O\DEanՇr4,Wyf*CvoI% ~a{SqKk@کO=[gW1)TSՓB7WX ~fSz8=Ҙϋk8<גρA'Lo)Tg Lwv K'3s![pw x_^_6uon Xz5ZހqEQ:)u7PЋ-ACAXe5 5$S(cgH4>@AۧCO>?"?6 \ƽ`3siޯa9Tzi6MvݮnBuFTOݼd9Y~iv/‹z/‹^~]LYJs]HኣPKV'b%2&AWQKIK6Nлlu1և0[w,K?OVF7{qIzVУh^- +L0ĹJӅI50%gC+FVTZlD Ow6r`jiC9LHbJE 8L$W(TI L0"@E[RS1&j1is B*Sa!^O怄"JQLHe.RSJ[RB`fR9BH7t[}BoLW5ٖ*[`U0QT\"f4 .- 82fD*yIJ"cH;E!ʀ+ц#P!KH,+U]LFhR&>hz%a/lw]覹};Q+| `mS#f][W\pae]_M?_m 1(@ R7tCl |:^-n҆.im܊iMZ홎և/z8leɯHG{sSLV2ZB>h( vh*w]v ǚfSyvΤ2ܺyQ{i_z֜u[OfMvhNe9/r3lOWoG/S ˻!If|;@ Rr7mtfzDZ#HOteX|vw6'\ QeT|4LTͯ:H|^-M73RLPz 3rn7׷O\.vjmYr1ժEXX^[\KvD F+#2)KDT qSe=dR[>y6u'W@?m 1LsL2$GxV'I%H/٩d@YOYķGGvX!Y^Q @˕ɔE*3ARViݖ|޵SoG׊#v-['jLVseUӢ@[9WN7Wj٥=0:$r $$:f }Pߕ˧|WΝry<==&u @ASZ̬LI,Bsorl&<޷S&1=# uOuzG_c.SCؿ_ϒjcFR@Ϟw|W}u)Of2u41ftj] %?Zl\/]M#NlPƋͪܵ")r%+mJ.&QߔHA@N{n5A`7GeMWL"ϖzi;<:ט[@.UjMn񵹱u//_;~ZlZ31б*wF0MlE\vAc(d[ +*/ۋ{Y ފP;nRs]MVE˾dB\`OtP vCvfD%!>x'ehBفO1|9bd$٪E7G$`pE}Oe.]DP`s;vI '~m6rcDU_}AS܋8O]#MOٻFrkW/A<be[eI^ҿd٢XRrnwO}&?Ϝ?N~{{Jr }e2n9\qNRc(m^X c&gsE2">2S(! v>qqzV/ J@ EBH(`X3[n6 a|1em%g?:15ʸ!/4ヶ3o" 3{X_\2QSÅ1`KN L1I̹- J%& MrdK(Q[8}*H-[f:+hʛ*,>q"[N!81QdǬdAfѾ,Id7e24M,"y-n Uߛ$d7Pw#H$pQ"~r2RiG#X@S(N٨B2֙.1hkd/|CĪ[̙vÜXU_JѼwϺIrbƝ}}C]- [S^iL]xzB8ڈ3Μ4lM)5,| ?*u|Hl?nlqHXFۋb^}ɪrԚkjfK#8IBt Xc k@⠀--EY˖n-CDN2mݐQ`>bFI~IC4,@~(]!J^3\-Zw CǑ`Ie4;ΈLW@BdՀu^֌A>td[;D1%-4~b9W Ms$+2LMQq S>n!eIM!^al5'J%Nϥ|W"R!4΋Υ0D/[Bc|WOxyqX ,lqy\럋(]Uv,+( E۾!jyR8?ߋ?VVTf,X! i41*:ca1mo]+ƐI@0ҳCaWj fRI.,pYr*^!C&I!SKy=ࣶ"Q^Su,;BL%# UH˲ȨJuh X-/]sؒS2Y]vP ˌ.5' Y'|MNWɜ^5h1;]-?2*>qjVA4MxgcBb1;NMO^ZIєEҎW{IdڅQ"Ē:t|ӱ&#hf;$ gg#e嚝p,4H';_A XqT{3ʢL&L&f͐1TE^sW!Ƈ-_qi!Ug0 'Kb[cK"s.0LO50xd4_X.@$ X\@UesΖ/SgTYU_sg17=V[+ ]((WU)>J@H .̧"%QDV 7%->CC/0 f|8V.{-mUCǎ#C"b2faB Soؼ1^2a=a̐sY.0_Wk$N08VZ5f27=fXĶS^Ls~n\+%bg:(=R 4l;|mnPFH $+*S_ aKԮNWTyoQ;VF55XmYBKFA$[҇ܪm);ke29"kQ!6D"E!fה"K|.ߓTIE)mpLҧ<AF B¹ *,ժQ0VӚB.P9!8 `L`۱7v' 4ht5.UɄe|LSAqbuj~(7S1ch!5"lXsrp>x4@Ԣ`Z~(kJPb1Ցf|,NHPv)fur$5:BP!z 'B6$]N_LFAY6w=3Ncv 8ly#^r5xz'yjS Gx!X 5ϯhxZ9m-P_r#"ux—.F"sr_BjE4!B,?**HQ綃 ?$Jq0-(\/FI.0&Oa!0H˚}9R9bi|4۵{ݔz)IiX)@pdsXbaUXI) skׂҚ)_"FQ}8a\+ɩ%w1U_ mo1`2Y~+C{ Ɍ--j. Y,5UfaV̳Bhy}2[P} OTa9<rckK@3M$8E wε5dFYPOk!ۮ-9 ax^;ukVȝ!6?k6ր R컬 BHTIS~)mA$tVvSySrYqŭ:h @qyaXIjRi;2Lp-EJP>NF{a0\[xPì<+)֒ajKAJ a\oiN>qp<KHXɂSgWxήJ^̰T0O<'xua0Х5R R*Brib"t֮Їiwׅl8W߯&xR@5hMBVi LUGlU1[4;'g[u ¦*C<$v罋 l'un5o9{$VH UJ`==Ctg@JfT7?]l.S~o׊pݞ^;U1IMG*/U|>[Nx# (_6V5L&b5KU/۔{pZa*eM(h}hgTO+RszYM#Mطɕ($0ROa ފUӒn{ a$5{G'O^+>OV6gSb8om3H95<1`<`Q'~KgA.%$o|yo*C|$pACķ]b^ŇvMuj/emP(Qy9A.$FV>.2l4;FGkcTۜu2gM}+> EsP{4fɳ^;Hʐ7"Sʋ~NMdX+sP";#;]̙˩Bc>EFY˴}7>D& ZP !qZ( ,jkD_m]o_~ØKO P8K;#2gcUr|Z^ ׷©pXKs= HfظfK a2iܸ.}v~tv'^)r[/;djp+bH2EE>>hF^(a0g-zjnA'p(e^kI QgQNy#+Ia#֭ͬdĀUu-˂ Hp=HltCRjzD =x[.'|M}j1t챙~E/p 99ṳ"G-k#%iqz UW!ŮBA' (<B=7 Qr3$?:)1?]]yD.Zh =FWP9}|t9X"Sl__WW}P PPr<?TgfҜ#Z:%ҴIoХ0e1}_1Bplt\d`%iԄK}SxW _?"tmۥ.ƾ'HJ#)}p Ttb )b5ד[{>s_'hexmJ>?e^,ϳ?ucLs%YiJR>E#yrXb=`8VĩQ;;ʠ (vה`)@4ަ#x˜Q:4D!9D&,*_w왒)MKq}? !cMt&ۙe:>ó,>JL wBc "(`U1BcBc.A,A$WWaIg@"YO<ׄ`tD}VO.x٣أx` 3\2\$e(MITfGhpJ^J.e.ыAwBlC,Z۷b[ԬyO2.23 !]JYjpA L3fZ6E8KvNTXU!Tmm Ѡ QL$aR$ŀ"l]p]Okkjm_d6h9'TO:&y^~-BZ7`oǕxVxn»͚Vl1 AA(ogѿ"*UNJ^|Qz u*$Bwb#9"rJ;O˭EU|$ˋ&sS^'kyuI_)U%[HK3*萷ҒVj}`v\_n8Aξ,_0>7M{MxB$ e9P%:t@"4;w'V1c_H4yj's\1Ea6|%T5M0:G;}ԤڰGt'dAs;nkg&RH21#тC;Hw *KiT22ߴVcO/)GL5ǵSQ>,.R qwr$έ%˅dFBneW"^cǃiœ $hR>\nD˶ź9"`eYrU> rYm4F̫;端m)+h spqlv?Hx&τu2k?൨$F\LWTBJ=`VLjRdYILhCOdq$ձS=w]{ݬ,z} גL $(C7_X,g"bC=*\xY V !8j\|:c/PhDH %t\~Te2 ;HHj)ȸ.tݻ k//f$+&9IsFZHP؛'P`y 7EJ/ZfߖGK5)L\ p4Re1[IFSw`;Mά;Rw?yLtrxI2Q͕mh0zVd2QhCDsCKwΣ+xp  (BlA'&iv1X%.#0v įcN}a]-h, .+)d]jiQ!c1ѪɻH[ ,,>Y}W]0UqE #EkM]bנ}%,hiLp7h %6&"TT?U-T-bFg+\33ҏXCE(8(f|HBҔ=Նt ZzGTJ,Vϖ6Znnw!dr^^MnQo2ĎmQ_9Y[-q5,Y2e טadFr&6*y} VSj>'tUBpb9Ŋ1!c; w,J$k9JzDXݫ?-p u0th<\Ws% Al=InV\TQoءaO.[lT )W6fn BHr;2p[eݎ<2ZUL)hxlG pvHM_=Hu b|>S٪DDOw牑ꀚ=p1PaWL-nu 8v~+x-sLM cqF^J2TR u \y |L!MGC3l27]"8 v %Jl}/'sq5{t[©oXy'fѴ΁ wjt{nJ@ttؾ'؉ mX|n¬@e`^6j!ִ.m4.UyPRD|岊gkA-3/'Xr٣Ek~8h俺<s:}-&_dڝkx/XQccVӦ"+Slo1oeï z8u6٣mN nw\&SfJ$X ة%~Nwc YjnmF>V'<120R%S Ő$+ 2+Vc$N=RRpso_*u)f/=d ZD#M-M'~[cD)e N-j>~BqKVXC=}IfЦ Iu^khgt]_/n$ut ˎjOs8-UM,$Fczā=;A@TgnOk3i&hcw.uh>5ŁpnA^1y=2mw p-B˰>zg7I&p:(^Ûx"fc}+XqԬ9*4GlI#5 zOB6Ys)NOWwGxx` qVc8}_4G:.N`aI1mF)h7gL^+PbG9b)HTw,H -5H^|pF`I>.C$HEz~b˧3=O@p)U``>ڧwf`0s(<3ORuV[~>Om17u'k G,gԎy)h j?k\5oWR'9,&$=s~cH{U-B9C3zO(ի7߱ms̥R'bnFz}b>`1umWWx8I%&X!1sL23r{К69cU ^ dY&02OF3K{p{>ߠ+.8tvjp+U\3>bЌȞ %u)S:K!pzq1\g̨,^^=) m kъ'WM6FNv=}1{k@.ĚVn{C^}_)i>XN@6L0OfP|-&??}V\J]M-1[ܧnם_$#X]C A`ܤd?PNOf>&ڐ(aFKryF0V4/2$ mRf`)R,ҜK|ZqjюU }N ] >-b")V|&r`yST6.o{ Q8=7L Q&v$.*c\y"J~HBI݂BzҪUa{6@mm`9Pm%+z[>PD |N26[| &t|;%\ #y[w¹T{Vg*;OG>VK*^[ "Bx7u{sD G%"q>+q/P-kWжCHu N@u'x1&7Blwoo6ݕ*Wx)·li.29ZЎ@)t$P]u 0M62moETpqVSt4N|N Bᛃ#<"85zuX՛@>\rh蛪!TLt(9S-O-孺0%pO lnID-0ǬSb ER 2*~GP$Ha:tgǨ=8!VaX4 uSYMQ+[S=?;2*~(MCq[Pck&RtE[Q 1z G J`Aava;q,yXlv9"|V?6 )*y"w Ɖ`"#Pir`>lEލv`iͤɽ~A呭qʪ"Fji*]_"tH&DBD2Fa?̧ i0.p:. ,w׫rUp%hlEldq_NYjlJpS&b\\]<ֵ%-umkKum?]J:Ϗ` e7lSg "01"DžfXd*+mU* 7ΔD*ş>+ݼ_J;Bp۠9] |%afeJ2EBD9Ѫ}pgƨb%KSW-I)X3e3(KwY؆#%rp2FxJQ*2L$*˘ bzeǿKpy&Ks,(0^H F1cJp$7^w)hXztXy(;|L3-i%)|Y9"~YDp2\kۇ I4t%cb{V(M2Y*/S+& yaiNsAJ88:%Z nh¹&6\< FKV*byyðWE4ھ=aI5v\NZԨT ])DK\D ́@&< "@PԄ`4A`%yJ\$F̔iAK8 rUy(Œ*0 D*))RBT@|t`›-3u= I1Wd<"'jo!42!('H2wSPXD I좻*Bd}>H q-H'+͊-#`6cPG\5DD'6<1T|GK@AŕL\IOm){qBgzɭ_im/ $XKL25xeK-mwPҥ,K%e%c9<Z|nNJc|ݸ;&9ZrTiy*N,N4$Lj,9eY4R~4Tzm ,J4vy{Ү9*Pv^wqu E9PAGC87!m1E"e9Ȩ$IH9%i@:V;oݝhKF77}y@3`*8Q+`9܃{X>˿rNen*NR鸪 dxCѭw=R~&ag7nr#ϽyKYCf؝tXcXZaO;r9ؐ`RMsvǾ4v;Fmc(;G8e0SH#Hcs ۘU26fB`$9%G~ qꩴAx= KϛV tZ%QR)+Dƍ^c,ZEDEžoR .:K$C0cd($2^(vWSnd9ɹ.iM}[k*;$M>0?.T`9OeN@Fp|(r1Dh-d?MIġ.W of> G\8,p⁶$Llw`wĝ kOHTۜsʍt/Dw}5-?EjdK_FVto#l͠flrW3/=^6r*iOO?$ӯ3NaA2| };MmofUF}{8I'\̪@7 k,2|miJ8Y} Q~Ŝ }ݐqr?<|@-LhYJo.9%Vu`bڄmIkϼ)-H4dm.90*!?7VD+,1JX%o52E8dL\R+Qf-k.mzhQ?FO7DJI;( W   E`!I! aWLgQy nl 'Ph /ZYޱ5L72@E`2+"ksI=뤓&G3VOKfGX<Av76K)DP̃!a?6!*uknMy:92A5IFÑи7nYggvJ0Ͳy)#㮝Wbr@zNK1sd (:cХ0qݝi;a/wN\;1xXvV<jE7N/yqgaU;B ICx~M{?A6 Y,T[YF? U ћo&b[&0;yO R3(0 =(f֫],dK&(Eph"&,.FD ,$ìF R~,GQb]#XH,F6FKH:Ns2%@a4tDd |sV"RJphwukϴZTwΔo  ! A-tPQuJŭl 鼵~6)n)=ƾRہuߊlho`0a! 1rdmH',Q&Ӡ5ǂ+!Z0Fim[;V_CfϓF% *B) сCEav [) zP .\wz-ڡTmc_KRbKK3 b&::E2G2FhT5@fe$lv8q?\sYՀI%X~x-_Ib4#m][OW$:$JR&Z$(c0L2^( (WJdlezAP2d(Wc he7o}*_1= tmW KWUZl\:V FO S&qց'yϫ-Rîr{Eqsbp\_B!jEJty܏Ju^a#2X0pwC*rpIȻ֚ Vks@IiTGR֬,~eiBM%JdM ;/p?ǁ&"R#8TZ{E̍އ_荘V^7xqd"YrQc ?[o~" ;ΣO,2kݛGh/}Em|<8Fyפ|Tft& V bkG,;jJK 8^=j`3n-c6Dv.$(E~#wM$ ^^J@HHZT>)952>ɤ 7Z0>iyRҎ"YQMM >ڼs~GF1?*=O0\,'{}&5Ge:qͽgOr{V܅-i2iУJ)~h\G _zï$}fisg >K|UfSb^M[_]:y^~E;[󗞝eNcNeӸY/0iVgfAhVhgԻI1tvui8n4yl>Dwߟ`5b|{(@"af'M6o:s6"I(XN F2Wae ROw((GơA#Ʌ14r^dbb^3e!㹆h\*ȑI%o =[,wt8Nя{_Eu-ϭqX~ Ƃd0=8 L1Lv3/RV*'u9-;.Jm6v(SEv h݉6DҀ i Rĉ 1ܖkjTJ8bƠ U xt0bxLϼr!L8G@98aX;h}a>zexv1,PoqV'BюT)F|/#*0|{M;8JA{Dg>I,"1j4bN;(0NJэuʋ>XH5_.G'"}k1Hx"aBBFb&:i\i_x4"Uᒑ۸bac"|]dDcI;#tJDY%6̥4b\=5#|ܳMbZA5X? ~3.۲DmH]eʞ2Z5!(16̧\fp.ЇOEW6\7Sk$Sd)ٵ98=ؕ\js ѕ<%%wiӕ'GɫﰡEQxƆ?EP!8`:LJT%i52Nj{@)N޻N)h#COLg!dnϜ눵SQջ<XȽiSN~+=`Wajpwjd|4+ȩ3E;d܏ ؇3f"Ϋ2B9T{!{_s FHH_U%rg(̾hs1]Z^ ox,XyrT-ͻz vj;p]r\:daM."jdV`S-ÅTҪ/ھ8aU&^` D_gюt0ꋗs_)2X8$i+_wBj$>6I+jspB3q`It[m `38J {ꩮ~FRHߥ;28UHoѻtzzb2ߦ{K) [Mè:G@ + "ۊ׾MY(ΰQ\͏?ШO=:vUCG۵je@Xʪ k_ TrTRM2erlAmU5xv}ٖoAhםms@6ramص ZEV7\(f蘣|K#JV+(kKE#Cۯf?!#0kVSUͥWyNr} 7>V?i_URRi|ן_o&{ˋ'XqSu ݼrUŧ8mi̎4k* /ާԅz^AZV7>bϪz(H֞[Z2|=mٲec.]BB:c}6RC7j 5.fm̽:FsKz@:[8wW1ImMYWyx5?}a`G OpӴFVQ۟ G&xlWHjRVjo*(5 #d 3cz#Swb|{egRA2yNOsC)O rmW (]!n~7#u`v] XjTL[{லo[X=:wV®=?6eklm9iuI2mL>By ,DpXeoGE E*`>wRPV콝lO1U%w%?=N{'77Z)1v/miY`66# [Χٺ#[Pf I< zXNk}C&Acv48<1D˿S:\c,-Ek@5FP,isgk6nncF0OUi1"k^Ƭkj_& ۗ]Hg^GRWQbڨ!(j,u`l_tL-n\x`"lRpHIFhk \9II?\^9xz絛;z'193S&9=RFp@ω,'@x-俛KV r8ɮS.zWYʦd)lJig)o^0hH-̧:]Հ6d-$яS?9{| lshlԁ"G!\F4 L h@telΊ+ `ɠ1q)T{Hngʘ&!2Yc O<4%0-s'3o8Z;%"f x! &rHdl2L HktyO+x2NI>+2$<¡leJmDز2Jܑ1>ipBE'%_'cHn\G^v6Uk_ܛ,De@I M`wr4y{ Ml(ΉY 9w$2F /2 %(tqxdz oP=;aC Мǟ˔S'"(BcZI]E NI>82)Y ;@C&K=M@›DS@]xO/ú~x͑)6plX~4.s&!^ԴQx𑻒A2^KMQLQ$9>sD3:$SB(y%cF7!`LLugA2hI@6spl)f@^d}lM $p dʉ`j"\BL4rgFxqA2h¥@qh?Q(0,x5S:"HO%:0 Cq M2A'gCA #g:&)7haB PЀ#nKV\H1'W{\+_tNE# s~9?559g^Wot~M≛0yQ.RYnYrCԜѭoY)y1ȫ}-p˭ jRr^N"&+:4FjF18I]NnuYv'o|_ҌLdF3YF9M0o!B8v)Li j] [Ǿg!}oeH"_@юBByɓ֢ZRbJ y2UB4w[ fχ5ˋj[]x㝋zCG}LJcDo9mq5x|q5C00_{ZvW;߿)jхX ZϬ4H]0.)Uˤ(^j Lg+ PL+i.[E5͵ @aV_i+0&j-',;ðar:r0DstN'!6ÆADk83eC%ѩ-K)Dͳ$KnWm-W{̝6ިFߛoT,|mqƣZiAr?].P}FרYpb: wm'd-l!iuTکeeTF :6Aۢ2}%.Q|>=[SwU+g#g~$ +(1;E}p)ػ+4IȸAy6=VM+=d)$,4an˱9/)>ɄԜK1|75U(hN׍|ԈZ7Kk슝7LPHK;ouQӧhYv%٤mR.$E͜Q%tSM7oƭVS7wӳV#(PL>`m| )h~khd?m\̚{׋dՖ; Nј܄SsoFHus񁷮B4?I@}8ZqZ@dSN.feϗ)Y ˛cgn6gΖ/ٖS,FZ*!-UVy#HB3lsޚÇ[$7osg1 YeO1̋dfYGQ-s~S[-̩_11 0S`L9M*欘)ӧ]N痳?$ qIA>N2j.c}! yRߛ۷Hͫ= 3tjHj8k3hhb;f`~EЈ55g)˞|M2[瞮n@D z7~xpdٍ&Z$Q#fNY(<4sRq2fv/(~Ő6q𹑁VG OdGW;!}\ԕfˍ)Fc@Hg'B]]B֚ ؠ+88x}rI%W!xɝp'c#_r1Ř4Cw룗VSA5*+8cESzuhSIj`۠N0ŝC9 5zI¾^~YHA{F7=xCW,zIJdHAhLvzܰ!w˳7!'rL֐5:p{ SHDBd'2G"\z-E""D2=71JteOB1KMY?{Ǎ׍@>o7&3<-[-=~fդ$"6|lj>u=j4m[FP4Q@.jH~׼ٜlȍH>~8nWg|0`r 99*s#əg 23.>Fp*5USJEx[v4r0}g'PE3BPC~~e !<}׻KZZʛ! a ΒҔnmUb+QLOvŴ ->yR@ɸ .-0 nz|ʞ㞸>yq?QUϽvgOd`9w_qiR{mow{%OiT2uݛh ׀Z%T h5eN'R=y;=TXqa$z{a; $0]OĹ5Bhidoy-MM!TܚNss~:t_#څRKE겧5)RA ٛQ)K\hKPdJvt T Fe5. Q;^@9FDNW?!hO_^,( r*޵ΆhMGbz>/AIN vy`7Q bgٟ F)`l{z"2O[p"`sM@ YOTvM}n/sN CʆİqG\˗%5f5g$XGPNv-ҙϾT#<䂗&XU?X{.BĬVp[mU48S3{>rayzgB9m;ܑy=lC?ʝTרc8dHr&K &{t>6:[Q#Pqn.SB&rDnD(hDJ|oGۨ[̦C܌iQ jܒh~F?\Sor-Ů0.lҤϯa2⼇QB6&k %w>bEӏ>sԓ]KyG ThQh wSI 9)kWC&!^rs{3&5Tg7wR6Xhkr?MQ7l~3]?w;. {Kg$Lv]Wceޏ34+9L8ڵ,+GVY+<(0#5»ي|n3gI=dGn /#(8WXW0gͳBYiq;ɹ) qMW?p/OqɝƊQecJ&b&*]|P׷zc}ߨ:2h:HvfHr晢2ʑ- 5}*T7F>"#?ȂjN,Ѹ#-o3bdGqkO9Ի{<ɧ2vloU8y༇+wv~ԑ?:z̳ S`ƦᣅC͍$>3ǁ-;x|z~%*l.IRab{Mn};5iWzj-^%xZllF,Ჹl\6vGT ^FVג{fp<#d82/F^/8{jj~F }LM}DxYFgG(MDcHx`` kdnJWh4_ˋ6cnhwe~ey+ۡpqyG =Ѝ}λQ/9ң7Ez"~thG!uyl4y8E٭nv>& h7ҹiv[jL*^j\ fxg~(DZ0 T%m51:WhŽ^0,>Փ8OS/Q!ձ`%#!;*(N׳TO>D7ΣKfLy;UGWXr,fY=͆x{oWG/&y8/޼~ T71gY>?mz3^_ u?.kz t;o7;K2n?!2|J"-|a{o՘g9)TGgΜ3%+p8an͗HgDV$Sb(^FP&'@R)zHAH;@r.5N؏, Ȼj>rRVg0ybY^` R uR2dk&{m'йB SqĽ#Pi *]\KGX"Xmen=!Dev)rIj]@:4(zo(YTfbl|mYV!b@ё$|DU-Bƶ>Uq}BuJ/IsW"z>Zk5dUBԪC Tb߇1P2%M )ߋ;%kJq"hD[ 0h=8l~nIF Bj`"IԄ CoK{0UjYrGiVbC7ƆjeeY弎0l7Ez_i3Cƙ}ӕ fsmZ-0'^ebk|J")0UD[h2S75V^. ƻ[zbW (]jKA"ՔC[7G.0o50jsK:5jKP 4d2-8jPYc{,Mx 5rܞRgRoA7堧+&`hvr.m#G["11?xLQ%ĠXnA8`SĬ!1;4_!wIdK8뤭w >$> ڨlo@..IVL ѐYJKM:>j#.YfFS]y6 LFBC؍$>3x F=wA6^rɞPb,u nP|. !Dz^$YUM33E4*A#4UE7ܚNss}PI+pBB\iE`F\l3gVt|IA|?~fbY4nH/EӜQ+C S5q+ٿN ڧݪ~>Z:|U߷!ERC%4P$4i4i4[Efk+k_;cfNs͝GGF h3k̜}3oif z ΢p~&-([(J&-g1~E;qI4kLc64rAb@HQ2bj` wp!VݛH>}y~QGyqgn0@\g=Hi'Z0έ1$I&$ Bsٔ :N坷ܿ1\A@+VSG<z~1v}}$ʿ^3 b ٍEQMJ 2A+CSUx!#~d=TC^7 0&lxQ{EO{|$.%أ@jOL͑C##V/4]"VQj͢JpI1g yddTkj"/Qч?ŭ6!k;ƺ{yr<9fЙ%I5[P2čC3FZ1W*d1,F ݴ_V!>LpWߌy6\=8M3T]FVx Y`%—.96ハ\|>q{[Ths 6qz%Iwz7X`Rɣ jZ%g5owT馶;rPOi ] -tW%D *K"Ý&mڊ/ ѐˈ!Yc%6$;w6`ݹkhtY`Cݺs7gߢ\Rq끇fi1φ18S8FYY 6ܿ03pGy47ꦶyYÿ6Ln=HHS.1N㌺Q JjDa8(%Sl_8;Vs);ge"s&AM(FrۗS,$M%Am}NN ?Tw El8Z 4Xt:fP:;7jnBcUc0$G ,9H᜻?##}-G1Ȑ3`9dɓj&Q$@-9z%q-j(v* R,%6f.Fg/Mɲ$!=̤xelN5i4yXFC#Gb2VJk8F ܵq%cmZj̈cSEV S6"u /gHu,~搭yBN=hs8̼֯7'VeHv+7􌡨028iK3rTÆX}zM̡#_\5?f3xħpli[ݿ :ɘ/|ec}ԢPGBC4uܠ71φR4$dqDC~y":̭o7'Áڭ. { );:A*8w!ƙk 2j9:ϲXxK3Y{0g ̘7fi<4 88"㞽3|1،y6@"g`9e?T$g]:O/)jnt6Od(Zo c3.+kIXCto*\YÍ3"ECqE2c҆DH:0H=9ii4z`}I@.<GLB7}BGvǛ0/kFcGf"1݃jX ۨB lEKvfD ټlnvM [eVhB;*pB9HB7Bq/jCVU6]o~5xq{HN]VScqN:Z98U&q9)>ڨy-M1&V:EAzgxR4H]d7W9Ơ_)se.6 a2ksLhQqf\"@!N0w]LXԴ+ZLAj%8,ٲva =a<"{ m<"C^('3ȍ$$!80Ha+!`pnx/8c̳ N/1MU]"0Z?$9sMU7.P-t8zyk6{ا~hx4XCNՎ@f)OY'tRj\fdZ1eF8H{uyFPu1+(Ud%vK&Mg%?{mzdյ8XRK0"hlA+4:]9G5;FRw܌y>vhjېlF+k|} -\6HʔpKݕ _}))i|ru!y0&՛`KS4"ac0x5:S$"6skz_^?PoV7>SbԾT5M΁>QZ#S᐀C".)1 `35E}I+4ؓ(vX7̭YwC][{S-Z|ҟK$gꇫwS cPn锋)nȇQ(5f] p꼫|͉zN,FKjBc̳!DʦhbktUz['$g9T$.6qQIQ̢ǘG8=gj FN֣$gXGJ|JdnalkmH /H]n`ؾX#v6FO?$4~$%!5!%l#mq83]U]zW{u>-tII'U6%o,LժQ]'JB)Jҝ. 'AL}l;YFN.Zn$Gk-AH3NDLJCF(υ!+'-=WCVGNԤSzRūWhG7c Uה"e77nowԜK:>ye?w;o}r=g3fTAl:!_wz @WAp)ړ_Lgpnř{?|ӷ7ڋ'Ԝ᷊>9;{q?Jfŋ^ 6_!mn 1onnϮԦ=VD)^N.n8"jtE6:Z!bhBL: 1h=疠BȎ DGi-GйGRe`a it^)[x/uƿyޚ#?vQfoR}JTU"{pC@KIKTuOɚ'F/@8gahcPqNhUl0yUBZIkrN2BBH xp\ H %kiпD5L*Ld+)ZJi\ v+psviWv|<]`UY`;T(\bEw4gת)(ُIvěۄJ[;#G$KqUGm7A >s|e._p$RZFDP|ϗjg2jF{ỌlRF*6]˹ 倖g59T_IamC9wt%#D2M;YIvP6(k} N.WC\|Ԛ4” А$`ϸUA5y""U|8FZ65rY#F;kև^/Tl!ᕒe06,!4ժv(ǣගBpQ KS]Y[>=I{o鿼{f~Ŝ_ח$Ћ,ޗɓ"οz$aoxi'ХMv4}l<J8.r n_=gߪ cr_N͔('1Q)YR]*ҳ9RzE}p$8UD>c:[D5jS7Ls-V9 aO<1FCb!օFt\{BDةy?B]/H}R8ޠhDjݎ""TÃJ'?ey7&+q["$1|!ȧ*)|TeΉ>9ѧVI] P 1T{2Y0o55(B_l?7>IR^mIMzv0, J{=TTs~.떕>Ow?1yJ4jє40|>n^^ʐݤ4L;hM7x0`ɡ'$6*?qc (yh͍_6'^jyL`W(S\9, )4̔k) ;3xj{z [ZE+fHΔړY 9~N)AԴC-1\w괛4D<#vp|Y%BOn 4Et!=9+,DTi\gKXwn15yFq)RqX\آd6bNeEQDjj *ѕ*QE[BîR\ЇmWaT$K͕2ScD\o ࣝPy&cIψ=b(i\1PCk#ʌN{ݣuϠ󣞮Ĭ}JA)D/81̳Aνyu#sқT*N'?f^R/CnPZ:U Sh5MЄK/(* QVvʂsT[^MqZWO*ӜSɼ*}o,\pg}%a<'$8ƎU4Ĕ*BYJDC8H8DN7XvܱpbɱBgL t<ʓoXQ}Bhr}@;! $CSTFzYqΥD){fhW$%n.`<c8*~:t겇ps8 wtä^/C%kp2O'gl/oל͞7p~#L^56?&\Кi}%Q3>p-%T{~NA:.%]Xr*3{SjoзRiC5G*I>)$;Q)݁C;(꘠KT{5IJ4w5n31&2hZON駂vBSҲ)ehW i밾n "7{>_?w|Jn2i=a@2JxVvQ16B!#XKA?.$zE`@-f} [Qn'}urތv?\B- _ǯwPjy4|;dO؅x,sVCPJL(kFU]yH:/xDƸؿx3kRvF 0 & )D@bҠ*i0e4S@5SPQ-.IJDZ( M9٢5/cp l3d"ږIAEq|q4!c3;lt] [gS.n?c@A\B@ D!CnG8^TQ:$-,Dm(TP= Y0enD*"2,؀0ijG JOjOM$"\uD!9cTǨ4` 0Y #wrMڕ.h9 O{imThqhKq=O-ѸC;F˵".GmϠJR#=:SJCn=D@,t$΂(q95!dHf :J:Qö]؍F;oP viF{c2( #ZDMJbciiIzԈv$FAҰ6ʨ K^Jp>eDC )hP1F#\D\T)jBF)VvFn* P#!nIPK ͸ѱK`G +DChtեetA N" IҬCIjDGwcWjv#q4\ڍT (GH(EG vG̚(S HXDYeЅJjB^BÍFig#7϶P}Z(SZ^I~Tȿ~\ʐn M\͗`$BI3 { 7f`omm@B>FeSԿI{iŋ+ueHFv4`2YrNh$7\4ˆKfd\;®c?hҘG\.j?¸%l/fd,~FGY-B0"^z*PQ^MH[|d'dpI|]#6x,.YLlĠےzt(8 :%Ԃ:c4frlf |evP)vd2h`B=xhfʥaeCAԫ4yAe6J+˔#>*Hac=AUl6\r՛8`C]Z 9"[AebeBZeݎƼ- R]h.6 : j%RvCKؖ5YcSS)K}{3Z'f?qv6 ̓mٔ(~=o˕ءǵoO(:zom|f:8nm3wl},0TTFWVAmiǔzB>)9%P@`TgZx/T.}9waKhI2]*U+ԺrJhIgבgA%d4oPuPþ$n:z# K@L|͝-59M r-py;˴ygZw&f|_d)TߗF#,#,sl@@]:'IgA3XOI`@Q dӒYj LRa4&DɿAu1 4BÆfXρ._8' XG3P 850j%m`r'F[-|Yw(pz5v?0ڿXAte0!ڜRw0OiIu+NfIݑ4S]:uR߹ICIyv~YonЗt;*7NOLFJe$"#?WVN%Z_⧴j;!K;Nh?rS-Ih` Ԛ$0\+%=nns6QK搐@9CwAwݎǂ;Ze0]غLL*Zt#}Di8׉3Mm`oS_we g;^#/ %/M8 E|YmSʊXQ_y 9=mg .; oG/ء%F ϙTJ+j%=[+ HkbiCXcç9d nX.眎˔0m.7W?Y<|eor.Vڞ!-=ֹ F('MD y4|겿,7p!>Y\_l_~$?EI~ÛJ_h|]$g}Kn?wIc][e(.?&ٓ$FmJKHlN7zjZxv !^]=15YJwqzCI8W>ﯓ,c?N&FWq4=1V{=Ԗ ^,ͨg8P.s9VH: ΰ?WZ0'sy @4 [h0l3veopVK:$K89#x Kf%RcS'2 q&r]N$qs肾rZ5Hɔ_!qGu`@+7f!:k`k99sʁ1+&JDjѢE<)w^ ZV-efLf4D/Y 8$c6u6<N)'v+v>T;c,Cga1Ya"=`ðcI2Yfi(4yJs=DZLp1'[l(JPccqa%W&2ѣL'PqQ2^FkTmY[VS5V"k˔BUsf1cՁ*6#4ŮNĉj,q.,aH +aEisr1J_ހţ~p=T ;W ݓ_n]˃ ,n_KɮX\*LvGåD(-*m%!4HIT7*LBai\+b# 2-I3dR _]F8w7ӄ{kk,,w >w{rdym}Ni՗yn:^CNjx:Ցn+jaeR:n5L*s0VaʟaʟV+]3"|rֶ6`"h+{OmBTx#uD(MV/*J*}S3e:h|Sl`Yx'ח6AUIBɆ(8fOlCˎ{b!qY!M-Y3JA\𳙔9=G[^L3Mg IRH-Ow:~J 5̼ 1]q.< -, ֔7>{Xv1g҆돣k*@dkc|% 'r٧{& ?,?ܫ fa4V֪@l%Dq91 ;A mlSywq̤7 Lf-~SShV_MEIߙdg+!jNpbWq:tH`UzU )BP".z=Om'1' m2|,2|:T mCӈ1&pd?핈'c}x&9聀J73W4"A( A` G x:ggXplvZ#gil#7J ߟ+[0@eCeys=J`a$;DxU ]wK:(zzx.2RnPt ]Wv=;ȏKn!Elb@=O,7 }|ԮC" T#iqѕ N53ܤbM \ Tn6t "ա ЁJ/+DKJWy&Me$.swNeϡH?X~&uVWowV}n^^{XHalңX#)Ǝnk5V_de5???ɿ}eM .MOõ 3l?~3X/__>Y73CE=37Шїx~W<g]ӹ/NhE/P9R]n;{p=_nmqjruŴNN^5L~v8>{oe 3L /BҤBk[j<] *J43\SxU|b~}<9-Ζ*^LUrqtzŗ~۠wx~-7Wn}n?~2J҉/؉V=4shoxZj˶_nz9tV^p1Ȩ<~~xGxgwt80h񦌭qͣb2ͣ%c =X9ѝ6 O%h4 Zг/`K/y%Ͼd/ٗXIeh1gpDp<lp%r J2OUAN3eȴ[俵iIA;w]'".ˬቬ\CȚ_ "{ϬBkPO͛鳢ViUyt^8‰;|I-: ;[z'4j4y{NV(O|h+%< N:  3RhZGk}M3%S oK&Aőg_K}-|᠗}ɳ/y%klH`/{oR45@Wgl9FbbRb1kr(kPYԨC cjr66Yyc^z\h©Uf .X+ DDH!^L)%k'S(-"wk2EZ,FEt "6C)L1!x; 4Vt] x $DMXJ&xQ |H"-Umbzj0hob7VSNP|zrϊR8 1C|YK `kK,璵Nh 8(FF8-wSlɝ]/o ղMi,w}M ԮҚvH;@ZPt `8>,lM@+7I: 2{q\uz-<ܚW*6ަ*lϯ^\_%>Sݚ\1to} 7*Z,WqM2WE=y2oiZ*VU~7$vkLJϥfARZ.P<2YޏSX.`(zMײRϊp ʼna?Μpq5bC1_=z+kD.!\/k+懝PDOe!T! d y&C!,`1̛+IR0Em`BbaShY2gV P>PP VkJ~$QGV{C B4Fr1ƯOK*?+T8z KLgcHeԣ)CEi[֡N$Obl[vNo_^6|*h5ZM7L Ltufn>YJoa<{^݂wBS?"MyGEݾTս&i7g{ ƟOS 7h6x0`+ľko+ԵOI7ok?]g-}=nq7v7rnz2 rTRɉ%GRF gitA1F/*Oȹn{ŸqT~l AܠP t;,{D(5?VfkB QsƆpS4vX]1\(%2{NkɅ˭Xs=T["`3H:W!b|,e9eϥNzD!~6\h0mA9ʩbcB^qQk7 [hkȅk Z%uOdؼ SR 4\]qRU!uW1}-+jϬݦu3){& D- -$A[<杠|{m&n i{V1u:r62 ?_򷣓˥DK֒ ĸ(B2IpX8kCa3Q/1<(,T3!ȄTegZ gcxvs ,4ͣ?/ L\-N4FIylP;Z W34v$!߹6)]8U;8c[S RD7h#^S)ݚgLn\DT% T=U&+~[S RD7hc"`#74UBBs])(SB"*嚬%Y.ƓWAPY9{@ӽ%;1 0ך$.yW49`EDYכnpp]S\sQQ۱O_yuɷJl}nڞb>CEjDV;'\ t !о߶0=4: {K y;!`(f%58t%5y-D̀zT6,5 |o@HdD>|z TaBɝxqRk4kܺSuqbU 9l1fĻZRT eVOU<_.*G^uFaKeز)S8So*Kd%!XRs~\8!'CQU@8Qɫ|E&(wJJU& HJLHO,'6PcAIf8i9LQ agG6ݚy$z ~fTVZZ(j~PI5`l9 &S `Iih4ca+("B"4n j_֣/u?W``pN8Kqhu2m@GRh'K1t ϕ"=4WXXR䒠+&47".hKƌRFEZǝ}Pee2Q Am8sC~/WȈ6j֨U_0f]4xƻ٘<憝~rbxC Y%eDeI݁q:& a[GR)mU/"7P5u[ړm'kQ|kU wB{+<~<9WҲL׃+2XyrwWC>W3%Ht<˪m4x)HO8FH< &kH1Z-z#v<0.TP?M8~MV#L5ܓh,5)6n"A:i;` 1ydWI3mp_#ė982kA-t̼CA6ӀF)sE4i#1RÉ?( {75E)ʨ>ܰ ,a{oR4 tz %&7iFL:09s蔹 `0$0YLn³F0rv4e.c76IoT0{@qs@u Ԉf[q6i&ӳ0dPdK 1.P|y%@a`LAH\T?v~ԕ0ȇ4v>y&&'q*ML%&O/t$j7v6wzH NsW4,fTH+;ӢzWy5No`,2 vyc5!Ƣ,7]s!YWg"Unqgb.ۡooǃqͣxо.JjuyMw|smݠxG/zލg?aA Cf'M@fUg^͛k*Fyyh I] ]meٻ޸mlW  | E]hEDR7v;363C"LĖD<T]5XYJN87~?r%M >jvƳŧAJ4ֳp!ÀVJpiz_5bf7Xk)͐ʡr?(Y[I5F^\:˒9E)as47}Zmzwuz[J`LSDDcahr&BaĔ(~(ZT'h4zPoǰ Q/QVqsC R:_+XA -O Í;Tat6;a`DAKSÅ0WPð'PBP)%)6G.bHPWm* (=BxY]9J(,#RQBi܆M,V"0HО uR`$܂:p(1zߜ$3LXD(EV)KTaҶLV\TG(4ZEʮ9ȑ S^xUiUY!v ɵĎhrwV)/UWz$gJ/W\ )I {9vrW3=*acʃ_Q9@HQϚX_lsPEVoeM$M=((t#=h@R݁z>atOUh~}-n~"^WY‘b x:!hhɏm5E $?:OH'2 ȐaK VIJ%x7dZũҲ"eX f }LI3>.=w mu\x R3 QB+Ҥ4 YiHTE*LA[Rh 24jI,koWbWxc]<_m069|r`U/5r 6쪧 [hD{߼]+8OAlnDd;%6={g,ZWV[-1i9r fO)}-kPbNN;TWI-jD+ĶEh}5?-Mv!_8E8-C& vAubF<=P6Cpq*aW*RWz/za`/G.`ѥ ^t ? "b{H4bJJMzIU t' CEqXXe aFPJsS0-Y,hjdb hcz%: UH5=Tly̥Լ"rf3VQ5+-fT0]Uƽ7 7\1U cڇ}c^Vp.* "IeC%o|O 8WtGI2FfMc6kB4}#GD O&? yy?[|Z,ެpL0iԤHx `J.= MN%okn-mLYB,sq3$?K7[ݲ™9Ԕ1p ?Ykwѩ9xeX%8bʔ%[nޟ5lXMM*sbpof6I7ݨ`9iTͷbI-l,mMT=:?>i!U@+E^5'_MMAF0PodI4Zd6NJQokhZ;K/SbēDQDm>ۦ_،wD/OJ[|옾؛`ވj`1l0v)~ث1Rk\Q̛FΛK?wII[{k ֝1-Aqc~<ZݟGKݾ>趪홇wwڍa iY- sϊ4O0ˍi YŽswx/{懟rˋbyn; $@8)vV%L4LLibgXSXu<[P$gPJABjixƆe"weB)f0DT_V3zW aa,`>CXgtEҼf+C{}LvM?hK;O!>0$sU=w>!ќNfSbf#,Rt BdH0d!$gɀK7K!{ =л!!ݮd45L,]͂JS>h^4flOZThVZo'_ctv܍~tA'S]_r)\5,r"T)Z Oݿ@TLsh_ߪQv U-"ۯ'Hx|sJ5o=$yo zvC9Je~|{8MhPvjs)ʯi8-ZY ?Uh#Ҟߛgsr#-RGp*~qύ=W^e3 "!JgD\%+S2E%;|ja!-,FiΈBvr*9>c݋:5< A//$99tqEOoou JorjPSu@HtOP1j> fhlK~f^~QQa,/ɒN Pi>@  1ɖ6N?` uɬi eV<5@Ȼzj؆|'Z)7yW~mɻN1)s#)yvRN3Kni:(߻q|Ie4@=ie~M)-D iM*GkRmlIEKe{TQe(M*-.cM*-qkIu,[v^J\(4O FI,I*:Bj1$Z(pTX5|.%;'Nu \<s> ϶XcRI 00z蝣O|]B)ѱIA}tW?!g$ Q9"<9b((x.rHQAMs1sTsT`)DL#8!G9AIj$$l2&Q |ЕBR!<[â9ɠ@ϾBF !;;AM%h[DQJxmk4Z'9S u@)AI@q9zv>:$AEm}uJgVL3*]1+(f.oWrTǛήߣ֘EW~G8aJW0טyYI¹0R$ x y|sok%ԚsWO Rl1CX4:qu@G^򛷗x~zpg~.J41To' LBq91J+6X:AX8TjXI%DPgD:QU6Rl{00UQ,7l-"zYpS`_r duoO+- le3Z CAv$4vNX Sm(0 @pX@ T]!"ThСהpB9o=M[xǯ:.v==j':=j~w1{Qo]/֯n>}90Mo Xw˿º[{~[~˺ ߼]m{޻G`k,S"B[2mzzb7:+#S1YsJ/k䍥nYr0 ))IkQٻ ܏}ˈ nZ"Nm B%' k1Ưڌ}/(!eKJ0YZlb 4.P0w0/_@O_0iNzcaZavCEË́0B; T؏ӻ0Ũ1Uhڳ>=}abW?x!řT.O,Vy'˻gJ8_AC p&q՛fIvC.Y]X\.)(bg鞞gz{"MD30n|_/kG逹tYo]̾gf4q(7^O\ӷ̴X0B#qvӌϥ9q=92=[be!QfC"W`Uɜ6|gX)t;fM?J,$<+)<25#v 3w?y)r?ft5&qysƻ;ung1"ba)N6(ٓiLg\3L 26uT8y¡UY5 *L砠%]KI4)igi(JVQzP5kBC Nz!촊{Xzz?u5"L)<ԂlTg&e}~p![Bgc^`𨇓 G㨅8 yhZFe["৤XI\xۧ<ոL!a=f5lx<~0sGN %*-6J{D,rQfq0ϸiAk ::پWQG:`nGoR5$(:nudkj>7O&:Y}H)5?ѢԱ/W?r룷yѴz1 +Um5pcvBB䡞t;~>z79f` gr:(bW<дzZg*`7GAW Re0Kh T2F(A볚пS;'KhXŲkRH],j+g;2SNo@ҬtlM#[7_>䣁qftƛ!y!dx0^-χx0j`MZX&qKC0_ncufzg#Qؕw?ւ=9?:| T%,{7z3rFyY߄Ѥ; Rn%m-$d/E[B$FcH:dLU5r0s%JdJ,a sM"H )#XG$V!TE[A ukk|Z2NNm;mkκ0~Uٹ*U<%PmZTT*nq 0|[~k`7n>{my1MW8 al3TAVw:q%pQLbt.N(x1+//%$۴ܠ4[Gh *!ZHGRgh"ц <ŠKWKjQ >)wXd~gǥȅI*3S!&rIYxF7N/׵PY8m fM>8tjA'[FBsqAR/! R_BJ+1&8ۄNHŵH5&$Dh5t14"EXw 3N+V௡qY$[xkyuN0Xy.?#7W>n*^G"_;l#-$.{R|H"L\|T=z"0G⺇|_JR}?fPEqz fP^ i)F.FD֎0ƗI)ɖz뵇K/v& w~:ZW擅v.Q^GI }xt@%냁Z=u+g^5!Co+k2 NԸBz_~|%Vyo(z)%grjk[}7gW׍_MwWәYiwQY N@Zz& Z)c1,>e+b %۩I/F}7/\_|h6~b>Cxu6ggx}I+셺m̃4tslr6KǤSͶH^y Za&L2}z aٿus:% sUNi~,+7:6% uϕkz7M9x1mTnIC{ڂnDkٔbg&Ynгnu1(c:uۨNL:n-a!_ֲB!s]z7Q tJVRc6*qţN nu-`!_ֲ)N>ƹ,[M JM5ۨN&#rn{Dk!gݴDgbP*d[J 'Ғ*,+7mJ9D(s"ӫn0z&F8x~O6L I gy;fl~c9iBΑ$˙H0yI0y ʅ^5s(0l8c'~bEygꁴ13pc8w\Oe3\o $yw=c!uag,h>棬a8M` fkZMLs/?~)alڟ# +$sƦ&0Q!ξ‹Wu1h}l?ބ§dپ럞koOvI{)emP{k|fI)Ѫ=ך'|Zjg&?hdއuo0C'ὂ Mםn湆 ᢭#`4:&(3,*͸J"$UAP΃TQt08z4z6CYVW ޸= â 0 G_`Tmh"\\ `!0(fx-1]6Re_'?|4LtMmv¨ |]6ڠTSg ~Ʒ^ͦ ^`9:{gL*W/fmMk/ {Pm&ZuNJũ3 HM08尥(2kh\I%9B"N1/6hWX(!nLj-H RU NAxޏ狒t'W V\EʒAUɻXSS 9 9$彙f0t M{{J=*]"B hiI=D4.*=.bB; %c~vXVgH>Gbx MmDS0CL&'N!6ۙVl typFB,5B(e 9}yA KƑrdaGrsgʱ3%m'~&\>ؚ׽LMgQ߮2A[nDžٟM;ֆhŸ[ 59f DO sLvN8+1FSFZcJ UG4q!j"DIF&7"pR!%1`y0y]3 ƁuT)5!R%]=8V&|G0BZڄ9g-sK9^`Dh:s(/j((:QaZ6mT ,D\'tyapʹUJoD a `'5X1+XB$Sugdߏ?HnCq[6;F^BENHz̍^  0"s B {QW E|QfJosb?]\_PWһ̃zP`ԫ"GǾ鹒Jg>?6bl/(H/bs%rV!E Ü ,+ɻ%W=%{BM/aa8FeIiH&%QRҔ&HC~]]U]*bZi!6;YN}s `仑mK 9Q?zi^N۰e;E=jxa#0©ҙf A`:p\o%!j4E Ros9sQTr6")0`~P7\8^8]eNJאl^/]&.T2jA%\gOA'H9dqw[v+ QHweC PؗXt;2N. qo)Upu`0eYK8ghe}[#",Ըo&7 ̟/LY>BPi{kE qypnR S팀_@N/k,\DN \;kw)vޔկn8]ŞENFԶQQT.\HCU.E,ʶsɢZj0ajbksHKU\~}=[p~ y?qX3x99Ic΁ݚug$PVg+9InPw|'T{+B2)%?7k!:>b`S@0'pdWLŒiIޙSQC;#5YzA$Ҽ٥n E[-Kә74d;x@TRagus0s$RzvMG C h6ޏ J~wcs.~M)pr߇"O@Njk\Zn.WnuUw =aiI[؝+幉tnEh//"?p) 66HvϞW&XJ|!+;JzP 'Vf~3n&Y.FɍaYn(8 p.N R]OSlР)P64 K܃Yh8Ў[ŗ| "8F"f/S^8[r=(;kvI. u~b\{Ig[}?BH0v99vcj;1F(k N.@0BXZ3 |7-n^=c^GŘX [[13aGm͍B 6Rp*U'Cbs9zeBP@FLntC!"(-}=~x;QRiC,nU%.#p .|׬#6քXI ϖjA $ZjrliV J=T_pjvoYQ+D㛘]T1<f֦DН`kN4NB)CC7t-K ^̋{5Í0s(Si hCJiʩ[5ZDQ:ТʠT7EGb؂ )@؀;%m'[qr%+wyI؆I{kH.]9&-pK lYNպ-kZ)Rq# dm# XX e%1BC(!aGt`otpë́ȖW5WS9'g%f݉SkqZܚ:pXϠt>x1.(\oS&Uxԃ(F=7M|8'WtqԗF.4~JH3>/)X"f+_Or!ب}'l9B[C>#H~y Tp,(hG ;嶲ʖ>ǹmpRrŊnk9r~b%+U Bq-S"ze\]ٔf,Eܝ<2B8 ,Zr?VGc=hU53 o/ճY"([^\2D9Ow^(uBdw=+y/КՖYG1A$XM%.8}JȇͭyBht'CIq5ǐ$ty/]1}vts47aJP V փX l%?E_uG,`e̎T-ʻK @ӵ$զeno(F*J :bXA-4/"?77ېg8^i&mK0W^e]{M tλ ͜k'9>0[gAND~ p3lz%/︔j}XHɎ;A^(4Y7Q[*cf=G2Nl; !`Bo[iX "6ѝϜyeh҃.5j;z61/_ 1մ_76AWr\P&8Lzl\FQ%UwrnU#]ÒovLz,TH\IaQW+\H(+{Pֲd6@9Rv%WvxoyX h [p uH \h{_G DHpFd65T=7Jl3ZhAd6uԚJ _a$[(B@G1'4V~9x?ywDK)}jj1d:;d_ y'.YōT7@߼ 5uG' :7_/Eÿ%MWooo'NuDt(A~'+o_}ډ{X|@)ՁY§`/ѽyu$!a*b KVW5&"(](~_i'Xgn^G}-QB>o偷yjߣep-W!\O?~=qQ:r;Qw[7(3t3dp[lVF^/,oTHLG~5v)CGQˑ5M-2 ˭M]-`VE7;")Jamj (_f=HR扥ș[ï!0wD%zK$TEy&~;42sАCBs\bsHbӪ댟|7LZl0JeqOMOJ8#YrFn}w1s~$k1.YJ-Z=յ^iX0]7YdE<3s+=i->j{LQA:T{Apd= {FHА 00 &;(=;x[~EnF(9oj6A sTƜp9 fP {%K-M ֢tU,:>d jYr|J֎!fb(W2K}jUt Sb+uvGDYWgx%6/\bP1eLYyQ!s>Lz8xFt =!piZt0(MzkdN{;lhwoG<r%87@dsv$]{F+p%ڃ ΅jnV :mHDjZ1 ؎F އ=l:ݶWH~`38W&U,j6#s:`X⭸5?}P2_#+p qD S na9P(#%>"TZҹ̑y$)^~/IR<ΚdCK5&u\̡KK6F:3Cl6miFQ[>fj`ڭ^w(ۼ _zP ¢Fm6wLC |vVyQvS=BXjX$}s3t K.%Zg]jxZ#SD\^05>O SVC^/g,uVdo:ؼWNnF]ѡSϗ۟/ųoLw{Qo.XMt0}ꔵN%,}*gJ+#WeM"IG'Mӽsjh4)CT*5ۧ_k>#YzVDmiX/y8?D_סN.xntV#WͥyukOsʯ{~ùWv9p6p4w.sv0u<Q6=wưu- X2aRYUI'_:T#~W j>vz3ybX^~ǩ.IYPP]?A'zQՓfvng?l?xws_ȳxc,7mߧ൮K^={{_tT50wz>F(Mާ(Syj"uS@X޺X%䏴ٰ: D =!5/B'Eу//.yv-5.&sjَrF~cǺWzfPJl *xʨ)*^+^ZוjJSrEo2}_oqMPx }suwUHjSc<%I2 v|C饦ܭ>-gLW: , j S14yuXpM Z~B z_a^^R9/w}A9UQCz|1Z}Z黠}g* D'?M޽Q&+XIEXԦ@<Yt1&LVILE\7V ~ÆNõl3E B$_h6#շ''[zP%M*lN=KK(5^wT:ϕIWzaCOI]Vrka^@^Kfhmy0_Ԑ(ЅUk}8Cϱ` wpCA"wQ<<p%5m Eq9֝NFDzBbk20/rwI?hKQyGC{fr3TOچzxp'[EIj_nW{z3zpNxO~`=.O`SX! ۛޮӽxx\(LPADrWKiĜfG_RSΡ3:eaЬZ9:X=16ׯāJQKA#4S{^8{&9yc5ygf~ۈ*/2$tQfT(0%{1(zG)`wv!$-^Y^ ε0FT\;U]@*%2i2韊Ф Ar -XdmD=L5Dp4%&kh\l Ű@*ԡ ==QXFmX6_޾!Ud~KpvV__mre%-QZ1E 2Sa8Dq{X5a}a.u4#h;q ,ï# q%uyBU;zGƸ|$%^o](VBT%r=ǿ2UI)+_geEBo 9"N>/ZĮSbX#~ VdA `(`ɫe$!4Rrmxx:+W 8&qR#,)`Q+!gMj41Hʖ@>fF+(AKDg]SU&tVኖR]2͚?cN.jUiar>xĖ -BO<|702g ȳ1)Mxu $5;GùM6(i:OxSg`L1uƴi`_`U*Je\a^Rrj6``*ɭQGxnHI _5j|^?TZ;ЋNQ_kf*"e;AEd)0`(g{ K̾n1<Ȯ҃&xhn5PO{0(>땁;\:N%"FΔu)gmO#ekiz|rdKĚy@(%F(XtVdzu2mGVl9fخw$zo\an;"6֮]z&{qsmKZcU)@ Od Cjr]Yg?.wz 8J4iㅤ*ph Y̟6qou< dbZN@ za /L-<-!H/"wqD!VYGj5u_70i&}ۚz=_70@ iT#Ni$}|>;Z>V!"Jϫ{G4R|Gs b :Q%KUdYV"4ϨZZiP}]F(U&k1T Ń1Cl] # =u3UQOHE`@FS5gV(g.8Bt"%ˉ:I(bpZ偕\AeF; ڊṼ 7&)jÎ>|j $hs nWCL `y N `%UֳJ F1!x7kFQV1>5b)xeE@U?W<*\D#dX%n ?}]b|0m٩eΖ:[vڴer) `R2)QZWA\ L2Ue(X;PTŲw{:0DB2 D}[}?!yU*LEڞy)  K,KN)\U2E%h SCx)CtF߁Į(a*L: jd" 2˾PAs.x lp d0lQ\)f26+`S+c5J^g+9zH%Gie#b^]ܼe%_]5hnm_RӞ%$錉P " 8Լ•Hdʪzjx fۯ~;f /2B XVۣz8^@|%WN?{X~ͬס.7@ -vdDy~?eրVa[ B!d]?|Ѯ VSqRYRHT\<vO>/΂rs22a YtRO7ݗ_"^7O,qmE#;gtCk`FG"DWPN֠( =@S辈veφcUX@\^Vv5  OdeNx Biت_JQLdH+ۖ]"k5=CA!5M;@$V;g+J?j~ZQDC)kPS(qh%oEUCލCWOi$j(`o@3SOw+ 749\!EإRf b0ەtv +R$)Ib'1Ud'>W G*oeO*"[Tdqo!0٥rD?|6(F=2luuӶ&>3w{z)^^z;c7;{ ֈTϿ6q" @5[] XcY]seWN])?|v$+ɻV[uc/yt}s3"{c1t ]jiOS{+1k™{OۀXO|C=/-]]lj[s~g,fSW%sM "sQ{WqB_^>::v8zX1D=#W==׮dmۜu]SIa{S*E N eBK O5\=;YVPX`/,8GH3p8=JyY61\w^Jg0lJ~;fW0/҄8`2.X|[I90Ѣ^[{&*^0/-#aqM72,5%Q0ZY$5xyTQgLD/B26Db9&\YZKyᮯJX:)bȹ:\"@ VbUFxMqAS a68L7\2p,ʃAIXORp ]<Ͽ堆p{3IG ScP~`~ VdBA0]hq#ՌCeLu%B`E& A20%6@=04mbLM O(r ggs<9S;ݫkz5/K{N4m)# Nc&4ƨKYfL`c׬^!=Crd1KyϘ-#$ert5wg2OrY7,ۈ%&ݿ$ DHmޒtΗ;yL+>,`)ٹ}EO#86HY $%lǒ|Gf;eå[.)C bNQ[#n[ջ$c3=B="V&%k-Bc?>[*?+m p5aovP*?ƃ>VI(cvZob3x kFy*R:n*pZ |)̈́T !!7saUXe뽿}n[%jVUذ65$~ P!xG?X{ 5YJJ%25a”11 :I.°hVx_hJM. 7PC!o-&bRa9,ˠ-0+ɠY(.j%b>owqZ'F:8搁cN qid \jf%L"QBpo`2x8,d^w;WBB85eKbH>oqپW&p^dv]b ry S):4)Ag\./ɫLUɵ5}]DϤ҄,}yZq`ocEcY^ ݻ*IMU_^,ֲ9M8ҐV\ֆ]WѵⲂB>LݓW.7st'v + ) u s$ >I ʀ 3Ԁy%C" iZ;Q* VY(ͨ=[|7fYbFQ*aZ%Ĝ ˸RɞUb"˒SƐc' amI%aD4Wy?ٝ::Mt)|1d-0b3ԈQ ;[ PAxU灓 S$b1fT F@!b G10yۛKrnowk/?UMW=qv*]_{k\F|?ynIm*~~镇^݀CD6Ϝ7'$:_mILzU,߯kwfk6q2Cl?~J t֜Hڧ!w\L!$fJPQTHCҁ!Pqb ;R"'OR`"!m0$F 2V(h> Q9>~e傒?Ax 1a&cldX=xIXw\ֈjE''#$/z7~K7-cjȯwwFuϿrB J5Q}ZB3hdՅnQh\@'w&m8<ʴVmkMhvk!9X>nl$[& c|(ŷ}\B}!)q->&@7S%:&Dz@,QertjJ)X-TiM[XPh%rr擷ρrX̉`f[\,H:c1r!g ^B 1IcKKL5%sPWN'^=ә˃;Gs@S_HQB9Ds0n"A trhBlBS[y"hv{z-)ɝv0Pћv&4Ug,L1|v~Uk)7]ƒ9iE wJRK4WXl$A+&47BW1g{{:qa֖ؑzVbj֔q -1D YcWR`USI8BtV-kT m/aϵL|ZLjOyB+*&>cħ _er~NWHZKhu>$J c_,Rz2q[Z_;8n(P,ҁ[[L35 uX@80+zr*6N2M|`ϻEk;5fn_ 7i߄8a %nY4~B+̺βi?OP]Kf) @aV{K5X?aV "Z(OY*f/M-㎤$j7ڕߚ%Z/uz lDc& 0 ˟o&[`w/*_ v)hǦeοh 08ƣBY -5q]hQ|0RS`wOp J0䠒Dڄ>祆9b.]5LsTJFC=7$c?nsY®z1n0bі5glM+c\]\XǛ8oDiD< Yᾌ c{2= xoI_Wkf1o>-_ Sl7?D|P<ޖR'iyūnEyR+ZU I) J#!b5Y1EƔb#,$4ǧ?]B}1_qqq:ݨBBc\XuYNqwkW`VS\9ٯ:S,3Hg1g Y]Rsɼ1T)!@8it%%< bAG%AC HXBOkT5Z!||)W$~H|쫷Ԕ7UOeb bǒVZ]k:D q7BCA4^]O>uumfj5F&PO]uInllf+1T?K :0d!. O0SჰF"&kBIp*FhmY @9/y|>.0f<ŗ_TIGȐ Ř}wI5|hP~*nN"@ ?_~~{%;>,cO;JoT`ZX`c3cxM橔Ǒ!U":L0G%?NCt$XUoDbplXgVn{"xo[ ˔96B*I;VB 1ޥSXS"ƈAPTlU ;`BtEIFS5PCZa(y߇qi\+Ε4VFF6e`5gmTH 於&m7ְL'mRy/Bst&${Yr!?|%rF٢\KYpx7_͜NNn6{3NgKu~ OZBi@i \I^ O@.6>! u)Nx:HPRzF2 }C Ja!&sO_U ߍx?٨\4r2ߘeQa?7\!Dy[qQ6;BVˑZ_ Z.pftenNs/v!WDM󻘙b>Y|Qto x\YŬW~%njFbꏟc񎣆U|<G5?=.fppcXQQo:SP?B|l# m/8^iBß|H Pa1?WqkGD<3λqpa]W@A_xw8R<;X༫)ƕ}37s'v=X[1_%!opkt)xvGzlS(֣음۸x0I,pƯTYLP$'\գ8:-[#蔑b^eCR}l4%߁wx[KlГ&={G>{OŁ€D)^as(O+8?29iC8p&XpMyR1۫,B i n@Z+1qJma$0Q ('D@|2cqrEeDeXr휠jHUUlp5Fj'aY8erGT{$7-Qm:ujq74Hik25.^!:70w ʶWuN L&$9>sMcu١{e~kA꙰}wX其Fz'aV^8VVS}NU4`nS__t zTM`%=;EXYmHBrm!SjQ |y]J-g,iA0#Tc:~MkLNnajrV^0i5Ȣ5~$xQ|magg4Eߛn(,[B8&#z A[CZY+xK4ծ&ɞ>$VŨ7EFk!A;ޱ:LA0k9ozS diWmYxW6} ضٖŹgp1lV}$uvu#TMr6N -Vc%_nW(Ϊ=("(7/vq΂{ɥtBGD!FZhЖ# lJ$"4ʔ vi:\3N44YtISVrM(B[ >eo5e7c+Y>D BY O8o񞱈S}Cxm6fC0DKt6T+T^G6[fԵZggRrBVzZ–WdEz-a맄$U3!Ȋ * ɫtOC- sSS^?w=<#vhzE 7Ud3Ն\Z=8i5ҺH^>j Dq;歏fN[tw#.WoɼH U0wIƪR=ѓWEmiuoT4a>/?{9|,+m ~ @ 2PXs R;u:K͇I-ewi̥{o>gdT"цBy"\{bk,{yos&uś"5 W>]U) Ds{x\TB)63CP5G{п< %:̮/o6'ݠxFKm|i+M|~5qacߌ=\&2y0Ef݂MEdX{BHdD U0*1s&5/hsDVR&Qj4BM/-ߟ{|nGFb[/%7 [&vGr*Y/= 5> w<]&Zx½FQzDcaǂ*LS1E`= ^PYvTKv%Ps8gs[JgT /DDCuɯBLU˺X KȇN?1U*HrM7'F\`B/9R RjvXSbrH2\ ISv|ɭk5Jϧ<'ő aZZ-8wPUbD 8BPh_к1Պ{oVPX`)D!n8Ѩ еT~}ެk%Q4(DzZ G& <6b=p-E)Q@wbkrQzHraJ ILMEڇ!QђJ,ۛte>_.f&H6]kަۏ뇐_jSM﾿o5> \?ޒo?;p^}_޼>vqqCWg拵Lp)rlף)8 hFWfz9ιfs-A0z )4UܒCx9rhL z&?em `\C'w-nC5EL?;OpR"fD !xRfB#r ̴L䩜@[p{hK(]XgIQ,؂=_K Tl$g> ɔD䢤`aҗ䏖COrE %"o{l$mO9Ox7syq޼mBwQ)gI~a{ꀱr JgΏ?M ܉j́cpe#\ԾڑkCfG(€:40562fszO'k `)_%XR~} w/ VJ94 ")1Bg&yKR`ũk*'zgl:pxBk lJa!\p*9DORO\̧׳{KE"  QyQ_63;)5F*L{9$]E[P>= GQZw6co7.Kb6x$$wZj0qtcf$0Qn?K޿>3 E .vSQ"!O%₞Cwu& ]uEMۙ(?QU;fjg(RC>(G (u<:xWPj"KEG{tԜG 9{rJ\Z2Z,J&M=NTg QxQ(qD^(GX!#VB %(l(܁{)#EĞ8{`ڧui]%_pX,#/J-_QcdCζ_爥 ϗ\{|yxv;ȧ}ŭ_Uxo7~_ժZUQ-9rA]+xA8azXlüa3ĢpiU/W }rβA><_~|;,V[mb)KM& ag%sV5 ^êdNbgqEY}[|<[\4,ņ{ yuD6U>L 8v3V:z\4n28]٪n[qQ1B_NW&/wph7/su _"O/2yk;߷N}]QO[Nkt.++'V#d{WN#ѤUPucVTg_y6wҷYmy& 3R}վNDg?Vf1LGKz3bRMqHn %roF>f#bw'_1w5$m > a/!^k݇f ,nTΦN&gKN6צ."Nòk1 8*[o e8&۴kx+wf{zau3lWwyDMӽ8k>,ev`/|J{5ܚ14MPedŧz?Lu ޻>ҶvQ7 mT!znAR/'/ I&@ЄqP*recmV9fi^}}i5яPzJ ޹ʾm} &x{kvViNqq$JԔ%$U_etI2I{L^c_u|IJD˞S$H AG&WfGIrBI#&ɑi2{zޫ̅O;\Cc-zzGWSxROrIѮߖ'ɋV-/;W7D[O IQ壪zҥNEuk(Wɓ(ȈF=FTN](rN <,㰘qIr9<O>N v)?lrhIs$y4Gx7=q:‡EV۷)ڇ-RtEt&M\Y=n"|Ya9h\^V@$ HG¢pmXZ] m%R˳BIs?G5+5EBvsB0j.SD;%e$huclKd+/YP6  ~B &yͭonTf/_In-|Œa_kg IOÌiEXO{$ 4Dh|0g_C&0!t2zrL3Ѧ)4=%^ d8hd YWf~/mFCGB4@?FMFC}0 Lyq8ڂiT#aP;c{o9-C}$PB3FlC1j${9"*'솞 %$TS{yLPP*$ #@\ u ]! hqK:\匉kF splR>8:РY|M;AbyHb3[g9pjU |Mrhj7䘫&s|:g$儌T/;) ~zirMss#Qpr L C/5C?\<ngX!BmLL'dol>KE0>B.~xÅQmjq;\g$kXL2V!FƋVvAΡZXp^Z.KsY`{Ff?o߼>2ҟ0=5JRȱ@9m&Q"lqTNFMGqO$EW I *Ӯ֕OfB RWԟ51EZVoߜԗg? W5gokcw-";| _ f[[`}dX%!51J :-x]L ,4q4Z`z ^Ȗ9hԵ>OF}z˔fءq qr|hars7>-nP73$ymVR ?4I۸ǘ0a|㈠9DB뺇 l -#=\QeM|aJRY%/P) M#'Ç*QE^W5 $łe=,OWtΔgfZD*T(czR9{J ,MLy.4>F$ 12Jaa[ŷfZhH:"`YD#C8u5E",jtp!jtc. sb<!+a\RXnTr@1 )顀!A`',(Ibε\E )]sB̥r+%R񫭏Nʬ~-~ք.HìHF(d1QxI@S&yKR>`"]CdmкoȄRĻ˨ GtM6xOKGJAVs*h D3LcPE!R(`GU.ӑQ&d1P9Ih-Xwr#GRa(c^$@RkF)2fs-&6B9ԧT-}J *Ӷ߆K\iaDd> 0o|~8[m>c7aKia_kƭ&- X0#xtHE ha3+B$;KkνpܻX"IsgEϽ[ߛ.e[_۟{6.{m I^O9w%2.lػ߶-e&9;^2&֟xP;=FXBoXX*NjSG \F]$o?j!D+x~Q-[ԦۏZSvM5Zڃlu)*mdvQ3JK͘6um,?ux3ȭGjl6ú.~DJ3/Za`&J M2 A|܍6JfN&if3MS hv(g KM\G6Q?%m7c>8Wn[` 2}A1!Bg \I^n%QN׌\E!cU%;V*SxuQ̻UnMhHUNiT:Խ!8V*SxDJ3V?\ѻ5!-WZ:ɭz7!xTQĻo5T.2_Qѻ5!-WZ:/{7j}xTQĻ/ev֯ݚАh-s|MjbPIujc#ﶷ-`twޭvfUր\E&kV.أwAթ{M;L|kȻկʮݚАh-RkV ru1:uﱉw_X)ߝwwkBCZtJ-\nLDδ.ƚ؈^[UU[rS۞ya wDV.ƚ؈Rw3v*wk@CZtc; Uuu]]:9IKΫtY^ #tlgx ď/kuuxZ>kjjg}Pt=]}O]Y "]-HW Ҥ%Hm} HXW7 4j R]Fɵ.܊"3Ǭ]17j \1+hcr̍Z&r̚"嘻s)Ǘcbs+rZ13n]9f).̐1w9&-#.0]17k L13do17k ʣ13Iw617j ܫXř"ƺscn(s]crZř!.scnRřqBscn8GcfLEcnC191w9&-S/5=嘛Mż1w9&-APJ/,;u9VyD{1kÓ:]9f5N]1TER}~ lW2Po$/r4*E^)Q^p0fkY˹9ZyF\:URgE.=! &^~~jAPj! 0_h|(Iz1xjtt;LEr?=U:Qeכ_̮Jpg}6蝙ɿH~Ty,Y"̉`gDq[8&\f:歭!xx\Y26i&!޵L~u;5x`a]{'!.CX*lR*eڐT! xY%Vai8)ОEWx;P T""B ϸu4d1gK!5N)0dƹE+ Xkv,&$602D<Г /3^4!AL?* Hv!%,. yb _z R 4ՄP|j)Uż\ sVZn,`*;kN35!,e~B';yZ@h$#XʉDt q(`0Mkz ^Zw @)[-4( <1P2Ǎe\јBTz zvE ,"l&o@/Y|1qIHVLc2 (d O ŭ紥N^#W1/0o[|;q6 (T/o.sД鞻ǽϧ^t ss3>(RDO/x{.Nͮ!<8-IzUS͵wgߌp6qkeJrïIA ^mkށ">-or^QyxpP}}06G8p@M Kaݲe0Q$,T)Im<,OBLBaϗGMx\~y48'pqyeO~܍L? cx[^~\J{L^AzbWQ Nm)4M&~G}q{NK+hkyOEm+_W#P@Bh xPsbO\6\G 6A_e b>6֟h%ee&l `THR |MYխYcJJȇ)u-A07dj.}㻉5$M?yvQ(rb~U3X1EP3ym앏6EI-vCG¯)uHl|F&::ُ1A%a[pK; LF>g_Ey'5z~cGɹ?xz ClQ305o̍pղ?q5uɉu!3KQPČr;*|:,.z5,YT`lRZH)?oIUbI{t_f;}kеoت. p9ƺTʼ^XN$II:3)rf-$Q{a^mf\!f? =J5#nܥD5,ο|VT=?$Z.2i=6ws\PA~r;~E_N04}?lۘvqOjzY˵_VQK>B(ܜ,pc2gbԁY /_%Qq: >.RE9k2&'>'N௾;'clr DF_$J;\`%/,%CTU34XFY}ٗ,f>:m*"p)eO>r/v>obX 3 u]bASOWkٰYETIUz!p^ !uFldGůJLg(x2DuZf%`S <0|eސWaEdq!g +t * /<͂@B8 L`A =>H4\݆IZ-vP݆f ;mjF8 3#eE-D ع&GWDW=ѱ6)VeL9Dڬ(3G܈ F2boq XOBe\u;- cTM!`4Ѥ}?R ouR ͉^ى2240\8  yX-SI1pW&&}NP|_7) %h?LS9B$gS7nٵ1_VU39%We86 DRLfeae:>NgLV=,e:`(F(B۳|Y*S֏k՚ SsYgB!TX[ڍ^[axsJ709& zTgt'%*<°7ﶶ O7|?v.V\V~p|w׳'&!\5$&⹅Ԅ?~Rf뿽sIf%)-=0.'X|sywmxE>oБ/VgwIi ¼.g>E3uQo@kOY)o٦9f=ۜm4gϱPENb~Ӈ]ҎG՚G|=cœ(?*,R cg05OEu x T:x,A뿕 Lqe}s7b 8y.賉OF|̦qϮa: i6r_Ok|bQ+wWS|wY ߉!tfF6H?J,Ƞ&JioOg$ۅhd!*Bz?=eM?Hʬf:uJ.F~2rZZxV.)cw{I|'}h/Qò&/j^KtH<71K^TLXBw>5"(ٗȣpܫSVxv]|FK aģ2/fg.bQ1!ަb!Ռx?qLUTKi*]۞dnKk$ܸ, }h)RX~l&S3ZgZ:BEqKzggwvh IzQ /-}:L$ٕn$easE8ݝ߼g#T>=Hh6ds{4F 4[DY`ǡYfU5kII>펂;\B_FJS+*%ԆSrK۪Ck3#ڨ6SZ t8β7 % AS9X)-C@F`ѐ6FI&+ \a]3e0*!b-ʤa$0bV0d0gMζ L3ZȣF0wX}FűSx0+0psbgw):05Gfjڇ=%څmpjzZ0=2'^utbi+zOQmd!Iȸn0"8 DzFgs婥P*W8M`RRJ()i3^3a$t.ˍůV13g#8Ty,WTy43ID>j 'zϖ[*3F3MnCQF%g#_b򅽔 / b's:tq?)w6\ ˳S:*Z3tAIJqYɪq .ZAY</oTV RV|*n1Jݴ#RԈ\* g9A! DVۧc>ol{7^Ό{c É~^Zqnώu.յ.'%ԑo'GB(E{tZxrJqIICH;qg(@#'׃?ϐ5rņ1xϐO*_'srW}쥘,h9`ixwf=\䝿?vHy!9;$zf YHhBC"[,8YKFL[#d}Ά+Wj^-8RrlWWBZ7çYa灎zr0"i:xVW LW)+|pR+1fDJ-$66m"GE4ȣY뒓q8sbc՝eAq}t6'}ڠ7녲Dy6[58. CHƒ%cjMJX GA- R^"Q0b&ne_KGh27Y-˵AoHeRZ?73HvPlFeGbцM'>\vgWسNXm:Xe SnYQp>KrA²!,9tEy&OdAfW P`UPTV t}lWbLcrҢ ٩ee<Ârӓn6HdC,@zFΔ@{O^~@Uj0tn[旙y3)$M6Y?٤̯onWR${r͟\<|opBM.RN&aΜ" lf]]PV\4stm8%fH(/]?s19:$u~Vk9)M9$GU6eݏ T(p:(y2ibCTtQ& f2q]lS&NtV;l/VHF6¬&IIӤ3-՜ݤ+oY_5E #prIBX]#C$_k-64k>C % rOZ K&:?P&SZftA$7Bd&vd}!QUMaB[˸z12(!Ɛ=f];L'7 LV i|4E29vL@-Pm35tZbO-RUʹ@t;lI]LR2O+u;@# O}?N O[6<EN OGjx^Jl 3ڡ:4{\d̒ }(39c!V[$ h_0G7~{x#I鱯 I)Ί6;{qr,>s=AxHIOtۇ7I|.Si*4 #8zؙ21I`7Yփǡzwi_ݠ'n0جj`.<1!cѡ"}'|t$>L9Ic=u7HzḇvWNVT@D8SdT_ZIӄmiB^Rf B>#8-0}J ))- -b4-3!$$paq/XwO'վﺑ W$iFHJDc$,b@2:zN)8UhAC?o}7 3Maㅍ#l\oNta԰ЇȚ,GہPp%ȚpuݽejwfL#0emKSL1%Pu29.@ێ[#j3޸kpIb<ˋUt+g yX~ 1E_Ƀt&7B${P7@{''ݞ+F XubOВNZut6$fRN/(ͮb+9v +_D7Joe dHlqx1篲Xm;|"b dooyҾ Xj  ntQ?"pbpi4){uE!LŲN}5ޘiux&l(BPM9k&{g-v[/e}hke z a3ΎTaP+Po(ȐMY|1%D+dM!ZaSВ5P{ mK LƝ$K^",R >~].j_Ωd(e$Z9 mSƼY(SjiLII r@(,:3P!C Roi"lD]֪A6ix< rf!C"P8nO\}zk;ƲA3"i,s=11Zzbui&ʖC+wwULG׊5KGג)6NwNR.ܨh3]qLRLOۗ9~BeP+SͬLVZ&*rr]C'8{3AF>|8Lh$o_s#jNA0)h="8TAȞȢ;w ԣZ6?z@Bn 䤗Q# 6A 鬭GDHƖ:Va1uWsi:"syE9 ]&?ܛe_ŗ^@/h93>]bց.D?1LD[LQ},C0e2zgPHhB$T̖?ۍ#ΝPTbSvM.j9*1F=uCRSl"s`c2r0mPEUomBu ,_:6YHp#D4SHH]irjq\oJeD}[*0u+A.m,7:~Y-c:a%Q:9yS%M$!2lͳA[@jqHPLQKP$[^)J$2}P {-)㔖}^,>`>) ^7f}D`5TknaaVڵl$6d'z'Dc184eS'T3hPo]]VFeQv:Ml܂__Bq/}w~f(57g[EOjO>输TǛuNRt = /}7*XWK6z]{r߲Jqw\S?TOt$?u2ɋ~sj  Ў|Ƭ/A9Zً7S}[Yޅ_b͛tYHjF]fS?K8kkVOwh~ppp)r4U.N1Dk(Y[s(^8Z̷ R% xZ'>=OR/3ibCFQ/a-0\=16ERCpe)sX6ҀIKP*!J7uC餓5W:`#qӞ~wqZ>^f&4^qn]돋w'?2_rՐ|'g/(緿̯&pa^l3f 4_ (CvNBQ)8idΕRe!\NV,,IvtR)Kp@X9G9`q8]m++m8Kb}0/F'3sB#z$Q )kf [}HJ۰Ϣń آ^ꫮچBE8r+'6u+,T Xwl!N~"*ەDhL9Ɠ<ߕ1p!Lu=W~1)Oܧq|Xf,C uے^e e+Ȍa8eRCɇ =LPHbqW]>΃3]U@ _N5HvN+q&n2*g}/=RE< jz$_ꑐ IvbJMnE"% T&:4|p~nWj0+ kjqVU7Ne;x#^ yrjoW$쥟Ku#]a2w0s'I0x3x}n|wP#QҘ;4,?h:cY܏ôtlN=ZgZ,j0܀-\KES`J\CϹCUߚZ?96)&ِnVNJ4asԑnOg$L+صtk)?96)NŹ)$[SN9HE`YyZ5)?9S$譪T+Zj#kɟs$3 #-EUymzj7UR7*)Y_Z-JSHWzp]zqQܾJ/AE3;ka'"X^:W^ya1qz7:%`L+:D)I"])O?f&|06]=WSM N :=X͔V"_WʨZ@ٓCH2OJMongj #O J[ X1<[9;UJmHR o&åI׫%.g_~m&Hu?J,_?N cv6d5~lRTVoScaJxi^6@n$G o@ n.P Țm2Z7@gG 1OX&D,R'?~9nNOq 3%:gO%oVq|&ggT8v?iBdYĹ$խ$4!jkF)H%-2jcbC=lyOrg4BavHn9yj% ۚR1"0"e|TʞGg8΃Pw;L:82^a}a'Aac k90Dr,#~Va.~tuTO&5A3[3J@p,`>KK~jjb$)+I?|ha.lR-jm,7%3ƙOWo$ dPEĊ0qpzmX̱^'h5+N]MsqkF}xM˫ɩFuX` *J!buuv F9da0~>x[Õ׎9^pZ;RU(nKq]ʆ^vr7Zcwi}9x7 Et4Xr|pa2k\):}!3W>\߾4Z㺦7@Oq`XJ3; -өk| G0xDPSkU/WŎ9U="sȒ-]4FtTh6Ѵ] >.}vwTV~'cSeߣ%;6MiEf-_=JoQA0Iߴ 3pCJHJˬLn|#w8hc1L[j[~Ts*Y0Xem?NᠬP`\f$`#C j2 匁][}`|Y{~a&TvZ{2 g+~\|R!.:x`9ct)/>*!9^3UG?^3ݺ.jK5VtZ2Sx\p+\cu8*̚Zg7XF#u|Cyz|pcwmnNMu~Ty-ξcu}tnn;S\#$DKQfx< ,ןT7auJenA H!'Gz&:tq4t5t,eOHkb W蜯q%-b[-hX7i;nP>3@v\$B[nk[̈"82 gtC3%Q5H i:VR&">d=mHP - 0 q/u7@3VȺ{1A2%m_8Sػt ̗Ympu_|_.FQyhcjBlC$z¼[oecEk--/yڒ1>zZWpܧv࿟bjghP2i4md8P~oWa/ q1"/REF/p)m#3l9Uz/8`3,KoXX0ǼT vslSZ`t]ClYekv˦|gh!1ʥsjᠺ >IA{vQ]4Oksb-@2p9ECQԶhDբLIKGf-pTtTMc]_H H-DQ@ 5HhQ=W "xUy$XDJcӎkD ںx6X(Yi=XʳMS 9f4 :IҼM4}7 jD#ӕV " aNr7ϛ@eQeLoGƷ'o̴IE5%PEϪDX^j8W97F" "RJU{9::uCGh s|tcR{cTO) 6HM"e`, cKE48U͋%ipeG#S/N&˕)v|2is'pQa`8r&ZljXLxMpZ'x(3HT*qw;On/O Łߜf&W#{4pSH$BB>vuA 'LV2y۾Aj$Z3nקkd B(MXV=(+Ez%HdJV Hu!B9H YmB1Nh% x,P#tFQM'?%!Ӥ*Z!!Kwwzx;fW$K3{N w+&BOd._ Aܓ`uWs>|5F^k是\t23.:8\^Xh8cz1EG-#$yi ҄Mq'(AnDнe9΋ rkw s+5XP?#AF)Ύ!QCPx )`SUЈL3ݡ݅bY(2(B(Wj1ј& X(\& ;"23ׄDʫbu.ROb61#fifΦ bf P-GyCʄ~N.R´;-W]X훓 (''`怹YoN෸( tV[+3}LxAnni֖Oi*V's]K2nv# -Yj<+{M6 :D/y8| PZ9E;WIxog ?:cB TDR0luadtGnQHWyəakM,1ctFQ NetsԿ\إ,vb0UQŭ79pOW<$W֣ިm7\ўsu1P V}Iג]O]UzՐP,15AtpE^2 U j=&FSDS4RiLס$%/aZX7\p-"Ƀ p\APndV4#H'iFI*xKnb꬟q)YFJ^28uړ`vh) 22 2T'K+P&L_mRgтT&9y-[!!G%L2~lLOE‘!E+dZ$DV?vSF4=|I4 TXJY^1bAG+@jJ;z^#-#؄hѩk3_;osq&3 muqr xѮs[c\E\ `B#Lv8?7GCE 1>ZHy$ɢƛ Js=z$1ՠeMVj{kv#pT Fc+{]q#%8 4 〼H+pBh¶03Hm[vۘD`$[g8p:0^[#~F0K"FQlH`I0L23QڣAcC ;0Q 7*DP;W6n,VA+k$qAj䚐Q!Yiփ ˩Í0>S;c6kbIDD  ]:[0*Pߊ1%Ԇ(K)#5Il'#IFH-c],zVpߞx^1nG,]Mgn^ᮗPzFǨ~=L > !h|PC"pFNX[0}a@{Lm^0Rt0&<\ɨ2c0bG0\"2*bRa3R d&ߐUsGr,UL[פVY1v]k@nt.[r ASRрLVXC=+xiXAC`H+u<)=XVMm0ؾjH`mm 75lȚbKk<ݬ3hDrAYg$?8);㞞%mdxK#Jf{t!`:xSa#4DJ)VzO qbiDˆ W0]~-3#; bDOgF03 d2\XYα1Z11B4Rh"Xu1+mG<#WmJ m%,6MycQ=k@yYJԱA4t`^"/ 7Ch́gWmo5=7aG"Yݒ^j/ޑ}8"Mfp̸9Q~QpyhY"DB̆q"2nXszդvYsv|[g^-e>oM3{Gq?d>RQO[P&^ʏN~T+a ׻l?3s{ɻ'-rO2ӱ^>ebL% /]5+ɎgSuI,}n|c=t>[czG~[eM{hx*IQDYOȵ`ށ{\bԉ 7'eo>;p[<&O U'{pP.{pOaYɣk E-Orb><ގL F[gӽZ BQkWJ  p *:KOQ`j2H<)PLGD0.XTաu'¡d~\Wtm<{ce,]ęw?ey\| < 6dOyu%K'K'K'˺D\яyߺΫ xíĝ9K,vt;/W_e!Lhky* ``0'i06*-2e IQ?8%X疹ъ?"Q'U-xWtIhwxڨuZP9$=BMRj=z[eB5y{q.9f'/T+^ *wp6q%-7?|BC^Sr܎RIuhKgqMTna do޼i&26+R)eTNuEقkbXd]s--Ydf7\Uѣ+gErlA?`B l3,F.a\]3#;flE7 { KHJu?>HA 9+4gv Q֧whCBdPSu {`|>Zg,쁄?S:D 2O:DYO:2^r&HejuZRGr50%'dHo fYOj`nawwUAI$TR$w<[(|a; ??2ע ޥ=y.{͊oײ2C[?]w\iL]B[KS>.n _fQeI)5/ϙe6( Öܔ$*[jNw*]+,:y1?Iot5`@*3X}rj/Aty>$[:%umPw$gY|_st'Yl,T'\7K5RL|>@J$κɌQtDndSe0B#bP`7$|(8zTo1(m[mnjы%m@R!_Mw!OFG 96휩jcxojOxG61ݢv`ڱKNLTtЕ,u5lQ_P0[KIsRw 5Ү|/uc^ .93|3/c,'"$lt&/E~jtJKI2N_x6`:Li0]t\1Jt b4KG,HX !t~m%e9S)B]Ψg0.TMq`i`$*djcfב<[כo~ߛ"$ c"bҺCixTXI F7A)ڪ22TRF[,HD"ϟMƾ`Ip Wܛ&:2aVY [+s##\;1𡠖 8U=+V^xnaĤbB95K$w~q`R^EqlMyFLHg[\䀪*ήLD .;8XI*jn:ɘlTy \*h'g\b\o+3%ɕ<D1L :XP40!70QXh]t$KIh+9tLI%SELPy։bI26*SnADXY)PQ>`L[~IO ̆w,TMO4 Qq5?ٔ>DZ@;8bO*z}60yGq%X4u+ \PQ[)JHYP 3"eAD S ??KP.ПzʍQ.Fˋ`b[ׄ:U|kepcfϩO׫ጥq2E'@F0y΍%PBM-e+f`n޵,Mmv'CcS>XCuGSj") +0X >V(gLk!4'8W)hr$KAZ[ $Y ȆdaHGOwc xyqsp)j=\k\<_3˥+e"W.-'Ru}u*?Vrw ]{|-yfؘ!#} w , "D&o>1Db{?PIIsn & IMqмs)}cRGSCٚdZ3u. PLM iJ Tpi/s eqk X3(ʴo6b0`Ϧݖ 3Vefj>E&.Tr%?TΣ[+0RHp3d`Vz)#!ĶaVJ_kpuKJp'A ⬕ZJ)}c (. 랛`w4~ EK ਥzDzh dlZlEQ uڤ8`DD.0=OvP'a`NDf\[uxX%>XGWC%-iƧٚ~\| v#Ht,aN %E:[ZAϣ?.LASj&aEJD m/$xuz6Ԏ|l>j7X.OphxdzC'63\#՟c}6Rh8ˣo>#Z,a H JT kIޱǁ[#Ϲ E-u*Ѵ_öB)+S7߶=L̛F}Y>36u}6R=/s[##? S$]ƃq;>$H$M'`Dɛ4tz=_"wV&dU,foA9>|JB%H]>MUp֝ƃkBiJ8b ԤS}+w4(VDv }22'Y(IIUS̈́hk;jͮC \"ŕ3Ȁ@z C&,ib4Q֤b4I;VXm 5t|nF+oD"Hs0E} ' '"vC++S98t8VE9 #ڧsv-hɆs;S)ov pGc'9q n=H|\4F8c4ZMWQUlJ?>g\4yQͷ;Ж5ʝR=LkKq/mX}MAL赗p뚉*nQ?*( p8a|Q^]>xT-V&Nbw4nR:_OG[}% m2*>!\b(!Xaw?_F_'UϗqT -o8h5uA0v<۵NvXݍu:op+W\ۭhhD{U> XC@%`F[:8#"[~-w0cBQ{^S=p$pHa녖TWU׫a1>S€ٍ'5$6VԖ8Z[j. ܥI(~Snoa]%7=j'?<;zqk'~Ew0U0؞\5#HGun|cm_S*ja;%7/P֜j@}:˕RK@} _w;zK´z_F1VInZn{vo\ Sʷ_9˨1iT*k! L-O+{h?E%(238L!*HZ Z2L&Q'L pYO4dc(WwdWc@@%^d[Px:Pv͒`+ T]/A56qogG<=2:NL4i0)`R1qŤa sN2fiG`w dk~.w>EF}ƌl-r۳\6{`+u<`+1XV֒Ō)9YK@4gLlȚ%tjf+`=bp( Bd6;v2;QXxHCFA[`L\E#aӲCe߳cꎸcdҕ318)D"N 'u" ,:z fĔFjK#aϹ{əbV:XC_4wlƤEطofY90LZ@̠CK w.$pt .ʭ)^4]vm{ 2p@k$zO#c;䐹LyJ63 d M1gVAd'g27| ED /(YgK6!2YdÙeoφA8SV}twF/TǟV92f?O~\{Z;B]ߦ>=n&^݉>'f6oɞ8=¨s؇O;fw/FJؼc,֭xay{4IfXGּ]OC\<0Դxax? #5Z⢠ sƖ5\`a0@3C; & s|^LDbE>v]-P}ED$U %}zcsFk*oq\&Ŭ 4wW1A(54}7um:uaUKb#UJ[3wBT2z+kc6ܮi٩m٭d=K؝<Ҧ`LVAeB267rxp0 cSjm9ZJHX1I&j"'2$RcKp۲@گd@Ȁ8f})5#ؤGrSY.ɱ"d<$\&a$Sj P`5oǜ!\R<L$9DFz D켠M"XVe)qa( yOL/(% F]g: 5.tm.txv"$GxA &:V|R2:1 #Y.%pΒO !k΄p*fI 'jQgޠ`&83ogr:Q,(kw’9FY.rx@EOs`u9MRN$s @/K 9Ϩ@z8lm!pM |ztP1Dke1*I-"א 8#(Ʒ䰕 En7}RhS& AHg`9"!D^drHViͽΐ] SV]Ù|Dn;V/Bym,cђZ ƔGdI7IKFA`OpHph Q(B4JXm˩42c j[ִdr %tlG6cG ?n_][ĊEd2?Fw)ܽel3%j} >-̕+_m9 _byvuA|Y= ?>9~XB2ʿA\m諭8u]6R]]Gl9)!pWMNC8E{=$[w8Wy٣$0!9##诤Td0Vy]2ӸׁuxY-tjXFr͊6NA6"}Jm2yy3_QuߒG'G'G'G'u}P>%OؐDH^+™S e%U#ݫhXuoR_:+,ٷ>B$ K]\* QF) 6cT1aX+#.y1ӏv(z;ޯ# ʲwC|n-]E$F;":f'$A)ky-:Y!+Jh1əlMp c%)T e %xXct֎5#*k0Q-nXdž$mIKn.'Hs#..c۶5Fo|BimkIT*n+vrfDbRqO./34Z4:&#&8~۟:$+ڑ 4cG?NN>Ottu;NN1֜,oԝ'4Krg(f)YUtY)zޢRyuz|~ eOZ*=z, #L!qdF9U-;dAts{NXRx٘Ґk 9@J:_޹\cm3|P+z\scdF#aQZkGl4e-eFcrA)yf4"]G3و6dXrWp:z)Eܛ6e_QIhFRk:^զ5K_7ٷeG3sz6]}5؃]sS1Ɋwy9M/; d+jw'&E8Lp0 z"#o5g<cc:1d ^|$%9^NVGr?g_9ykZ]kn]Đ=h[x{9b?zyeh!ҶZ^^nɇ\=ŅAizs-dQ.<2.ˠ3 㨶#ǭk7M}B#Kp-gmB5#J!Czu!#xL=0LU2[rWlT@[.\PsrpE7gnge|e}>2\=ӂ;Z4ݝғSTXAOuA}bϖRzHVAxA <;1ϖl%bZXȋ)W:BG|v3A\ sp]q/k[y޽PĎ/BM,Id_cLcQ?1H ׊F"oiۨi]{j1kڃwH:Ok$HMIeSX(UZj?LJn09:i& {. z+kmHg+ ET~a[OuƷ[1[/ZM=/sqxc3\Z)X P D ^m۽hq 9FdKke86fLw/ڐcYГ8-Y s{)rr _{%{Tm]xm0+Ե)woEd)i#ރxKA&Ȋ`:Rjg% jkL!  x#=+wϹ"!8$׃kɫl2&)D3qeRM^Xdܥ]~Oǘ'C(Z6Yo*CIl~w\8\ͮϪ7|(Ng} 䫛O'vtw|wSu҄f̅`)fI5'tZ*CRFhɣ0+O'֧aQh]Y |}l$;6DNN#pMvV^2]L cVByጤ7!hZiQbۣCeϣ-=?4=CLpX+~stȖ&Jt^8rfl J2[thAGa9QWthnjYCʎp"r'mpݾ{y D\)ǂk& +؆m՞ᤍ31/i|WR+Fʮ>v@n4s:z'gm]51JLj2Ƀ.Z ,g eDK 7_#oЧHǯDS !yp4+!@7 veb'Lz=ڕCta\#,W|dR뜸WE˗y3j^ՠ{ w'J?ᛄm[>ы%dUwmrNU KA^$8@,M[i^Cʖc[d{<] ",~bUUICO}9sƮv~; W`:+t]Qy.~Qc5j~;Y*-).cZȤU~jխϠ]U45 WTh+QUN\KC׀kWR~ifZM(i5u;^0҂TljCV6p/tҡׄRB t~sK5`]ktV[hiXܒWu~яbLzZ37xr{}W|0Ւ h0k =$1 w"֡8d7),϶>b9 >kٹ[ս|Ol9Z-Gk̇q~Yü}(ȴ;xG3>~8oOc$rACa˗+ Hn *yȩߏa">:j9T=8l "8΋/C٫^K $/G8o eֲSlp"qukh`S\J_U[!,PUS}ik[>)TQ$W=f#0ΥA W}sEDڹ 8nG2΍C~-:^?Iʺ!~ >5:cnzN=knNi!MSJ}xYNA΄-u bN}ۄQ]-|4ҺC~-Swf>w,t0% SVBxUJ[o0U[T8~h2⾄^%`* :uf\zl*ܨ wp#M*^ =R|KZ`!ZrugܫR-Vh8lK^\Tj{ƞV,yZqG)A#h!|{mCTHSp[/]ͨYhF\qHS_W| 8JIj&0花 Qo=G. /:c$|o^h7{KH-%wV/=% xHindod\ "/r.WʤJ@LM>nCu)F\ho@ըc\HAso,FvO0F#&J0MLzr&ea"B]󉇇|"sRȥ=?IjkA8XCTiDM53kwnbfp注ݖyg43=!C;EL3&R6!v}1^*Rp?^#nB+u 5kiZ dAθT4_#+ՍwMbX*=DS5+rWe4UʚJHZ Pm45*iydtiZ :Eof`L>A.jR}7M>ۏ ?:}&aJjgBzemAH)!TdG|}4%(4}fmy#e䎔E'{F6[0w庽8lK"y -NHS ';K,VߏsmZOgŷ/Cr^_o81 *8"mBFȕ#돔E6nymo i2eo0dAK9U&HI]C@AuaUy#:7?M_GK h^`鹲j 0]3𞇋9ܛ 1}n;o$uIC희_ܔ1Lt\,K1J]ѬNNn$\}Ջeހ-O4OtR}i1_qgLeSQWy"کyFW u^w Xj<{ʞCuk OPc,cd]" E!T\T¶4$m! [:y\@ g CsQjmMwqKZ+b"[jÉ*Tj3f3ֹÁkH k[0p@KkP=clAHc&q.Jt0O]]Vj t#?iV=cla`1 }0!Rê#Z% M %A} HeC`k=~c p \3,dp, ;T#ӶZoKI^Ph%C ӫ.hPy';.SA&D  sۅ}VW[4w.NnY $?9$KZYJ )Q0A[,m ew/>I:26gd.dɫ)?kLGtk箵WWO/3H&Xcztm9oZ{+},5Λ}Qy9ltÅ\۫~*+ DfhKͻ* r;&O.R$ֹFcC)H8p"j3aу5f~cHX(t7XKɞ1޾c'Q3DIVP1М.(kTgT4̈́׊7׷usNDGKqUdw@۵S8dE.٩Z+̛yí6Hh*_ۿ/6+mW7_^qпGz#=ø0-qξd;u2p).3.<?P#wI+^;h%t&%8wk\ ! ([`}cbwFiCY]$h:o;ƲjW@5mАƝuBȂ H &mhgz|b)kQ$)Q(-z 0;"D1V!fmrJd"Bl+(Rc p  \0*tB꼇۳xJɺ =J^a5hCorM p_Oo- ʿm"=翳m{ٮwnCZ)] Ur@%3{n=yr]'b 7js r.礸4ҫu:3Ìnׯ/˱|?$wZ')gʗ N͡jP8h4ۛ3n&$tXdup.I-vd2 Q&6\2i GVy `٩ ]b2IYޢc[.C^J4iAQY%5\d'Njcev3{G)T 0jrڑ=clQ\OT)Cr ּFI-_F@0dRF6Ȯ1! JBL~z.N{PBgC4nx[uCiǓWÁ:1v:Pnjŏ:ޣϗ/Yr7xsvӃRU*ƜuZj* Ȟ1m .{bb]|i#%/GO o! %>.mki:U4a n[I@xY*gXknv2; Șҹ'GƓX<*EZT 5>lRL 7%9 XiWWp]_H#埴S%Js2$mOIAnZ{{^>'lfN>(Vl2+E]Zמ[8r*3Bj;gd8aem"Zj%:" 4*o-cx+}c=]jmtWN@:۾vT@udTŽMu=ST %5-F)R'ຍ^-mj|%"E齀 .g:6I#ke d5Yo8vv؇{ybRтmʗ"YfhJ潃Xԫ {Mhf2^F^Z2nVj~&RF )ZZx)NҖ "+60I~ft$UFp#p*xͭ~G܏̡V23p+%X~Bs J߼HP#6 %A8Oڋ^Z|l g{6WGpNa&~׍hn& iV,W8F$E?!p$R$;;h"]n.Q )1gjxbITD<0eReZL +ǚ!o+|aRֈjaj0IxfczǬ,Df2U6a\g7)fgz9>g虠ZgUvi~>Y|fu?};̛%Lz9:cчLq3%,J (՘=*=!+^BnX"0R>%,6hӍcZ*=lҝG 3./]_LUu[gUZ߭e'H*K^٩VD7$* O(5Ǣ]M=pMY.&\{'\S}]pM?kQ,z *ΑoNk&Vgh_0UR -^@`9i'ێ3w/SB &cd2J* L"̜lw(W7.zm\KNJݟqIi5P I@'Bڭ}v0݋[};\/')%iz74:#Klv+kM0<J{jaZK)*gҊ\br in+ b 'I$(,R B Z ܍u lQT@QH㕅nCn8&x|5؜3]fs4\#sϵ<'N(wY$Q ]LT @ef H#q,UQGT6T6^ͅR p=AGseqn7'@NxisFuNfw(|Qu VZn9O aJFubi5 =)ǹ1 'C)eD$~t`|RBl$qRH^\:/4 9Kˇjpg&EpIAt`] 1NU20jlS)$Dk2Z"\Û|5~kA,ԯethX2 z 3AVio`v1k>#%8]>Akw^M} N:qmu ۛݍ|ͭwIhS4s}gg|}cm滍={voolnwްn7~mw|O4Aw}[ #1'v'~7@ :?,3u?VldV?OYq%J"I ]^sdg45Cb4y6k:oL)NF@y;p*Y{? fx~uj@Hx*>QG;VwOlqޱ/`z1b_7tNE˓A}'l'j(>΋-J]>8ONwq{{ۿa˚K-1կ|Qo.4Lxʯ'J^t{֬~=~ͣNZl7E v' r'(%kt;Stu/MG%~; [zdjxT{Ӊn(KwQW߲nodOڳf|!qwgX _WN6v}P^/Gkluo{^t9HH6tNlawˣw^oD?Ógaҵn3ݑ0Rų~Nˁ\!{0mσ@g]1\;cʮAו}[ˎ{}M:Zk0^Hk4ԥZMxr<'Lpyatҝ늎lC "0_*]8A\nm/-+-[hKl6fZ!3Xá'4C=.Hy*Hy(R١&D(>hOruc$ņc'1u(&8zhKSC]Zukx>S`πBQJ>"X> Ym ;p4~}X)υRD&?NOkBaeN\qY!rJӧ3[>,y,y|Y<k֑0Zα:hpU: I8-v*uEnCDndh4R'3)DK8#"BdJr.ץGCWLr8i95$[# ;h2Z=EK~y-YFKђed-Y\dƚ7I&>7%%Ia@ ypC QRENfR,>O+2T,>OrHuȃ)i&*K@ze1R|.$إdQDnc݆ju6 QK$|Ʃ1hRGb#pΔ >-b :) NiƯ/ӸtNi\:sg.-v8k\kZ4ךZm8{ DZHX8W `.Xv>v!cy>SM?QEVz"Upy\>D HBA D Xj(Tw6ToBkp ?e,y?y?ϖybs*|F+,QApziF,"2B7ꁲ ⎺oQT ,\U;.aMi-?]` #")٩Md5\#H4ByLw?Kc7Kr*쾖nR篛HuVz*Y (ZބN L%1Gh]li{EFLČ&fkb3)ӗ2}Y*ӗ2}YL_[خF5k sb(7CgJGL7I0NP˙f  + e dZĥ 7»s&AIT) N@( .V%ϙ4HDqBgӒLb8 qWZAmI4y'T9KT9+WdB޺5kPlLyLSʼGݸX7α_}U6W5.6^lt'FR;O64 xӎkï_C 5CȬ(3zPDA6A2%1@bE *x.JXel{ƺ?Gib$(@[P2##҈:Ou#AQ҃@ߝ^eklQ'X= c95ĢDU@6 1<"aU3RBA$['Ho3z#:#"p+kl|ʾTc_:(caABfWڣ5o6&Pz3ue֞o>E*9Oݿ:HC~)YFV:Q=zF:~u/~Ѻ;|B1]Y3Kwe=ceA8DO=+h>z0Ч[YV|j1 ZXf̡!7S#:(INvX FMb&XCkZj7wTKǽbrxiE座Z#~"O}CO Wi-#p'dD2竣ՉGiz}ǫ^񻫫~ֻq(/Uptum] Ryy,#?ּ$,<_ >^ydM]һm5wNYC(㻙lqo^r9̙ -eʗA>>$<şO2Asŷ(%oևuYOaWbѝG@ט}Pl9hUrp Zrv59I7c}׎OS7KXlmgjn_͚-u)np~j^"|0׃bm.x\@ uRRz9MR5.%KӱZxcْu'bP{fuᄂ 3?nZ:H \BU'Ya7Bp~z;-BEZBS!~?n75߉8[\+JHo?2_v/,ا9oz`; Knn)su̗ 6XR>7S}5Hh\b#ȑv.ONV3*h lU*h=DHyzs(y> ~ |1|^F?775yi5k˹Z8)|W$>UfP^DG:Giz1M|o׏{-/"_^ &{nzȊεڴ&'l?+9$k|h`{CY$G0 &Иac͸1[.]h&!u?,ȱsz,3,׸ 4iglL` Krܛ{/4^XZmwپ hXމ[L,La5plޘˏ)N8q`ɗĵV'(o^Diׂ,},Fઊ+QC&Mur[]Be Ǩ|p"nG{᧼5]I2PZ/+G&&%1 Bk@8C@#E։Dn, c@.-I;J2Gq\ +:2x ,f͉=Y3NalL|Azs*/zkrՐa*24esE"c .c"r J 2@]L <+dqn%1$K"ؤzEIcf[EZ~S0VZF81[=J0jcPx6m#d3!,OH`A~ҟ00RkƦre`v_SGS9P AJɨ\C :!5)H:BAI!kP91r6kGeOP]f5 3qIXa,C׭HH%0dB7̭JE$Yز1EȺxSOIe6 l4Inؽ߃iPq-o}-oW׊N[zq~ٶu9w|w~-ǖ"jI{~ZU=Ť/RwJxUjD͵=ocm "K\&LS6ShG)b9fPVbZ/> `wŕj;adY3]c2)U mf~ rl6YG#tkmVn4lL ps0ډ".Rx|q9kmkնuZ߃eA] |]"NWrhm"C, *8TH gijK2320`}AfÂv?\6D9` .c)#lVnO:ʦqSwtҍn+ăD ` C汕]xm':Z%N0'g#v`:fdIS9$lh߂8Ifóq"Gcseڞg֩I<.G2*lqU K)fDu=LX׶}v^}R-@pDɞPS#2x3{6ʷT MT, T]I(D rx(H5 HQUV ȕm%r3R|dH5!I!5"&ni2=QR?6zlp o7U+B8!V/%oƧV{߶}jOzJOlM2Ѷ:k#fH'[GftCQhNzysGW1 !l=1^=ʎʴ=ϴB vfFYўļ͖]׆[1 }첦vYX1TmWUɾ6[JGֶ4bNpRyt~@b"΃`H IqrM4-%b-駃MO?<7JzCǕi{)3BZ?WT3Lׯ#w $FojFX qtؼ)TWd9*3C[hHLLr?ZL=e~IoXlϒ?{WGrJ/ ,ͼ"zxf޵F!OGlJQh]q+#諳9^X]Ԑg~;oetpy"s֔$Fjm6$:)zDp**i_@!Uʢ$c !<ґv(m16Y:tM㦌q ^3~}]Aq]dS3]sa9rҦv`VuϢ"H%|Rԝ Wh YKYA"IZK_FYTM~9}o38 ΑZ4R[ +1ctR|DHP81]JRlXQg7;[WTU^J5-m(+پMS+p/\{PJmĒ&. e> V܋sw퉤QkZ>'eC8T%=A'00:G$\@Wt-]FX)@+))#RP'/U;Ai@>Z!XsO(3.!aէ͡68ʨ2>b[߆(hܔm&7Xfrqjh2/ȩaT1FDtςWqF~=c\cwuDŁ }@B7(ʬ@=}-l]?O\34/"mrQQn#Dn(sy(M3<2ˍlb[S|e74GCӥ;oelR+Qt 4P(d.pWFQ5fF34G8 8N}0.c^Έ!ʈϟ &u4\W%pNuVQ=+^(Pk^ @\|qb}g-T2/rۛꬥvrq.r}8^pnrGԴ",|\`+nȨa}LK؜|Z_528,ຂ|h7n/6r,q-f>?Sg1-w2u˛uB$V-Ao6p2MrɸB{'ɣL'wEPdh1=[^Ry g0vB; NE;7H%3Au씏 \KA l&hg@g%jJdzYW],F= 5 Ӽ(p ˦}bsZ:ڜ.VywVr!}wdPVaf 3d=ۺs3f} y`3vRBQZUpvgx4EuHD:D:҂PlZk#vWJwf=(n[SK +;q^!UC!UC!UC!U bImAYp!0dT٪Ľh'ů6B`n0Ikwx_vwWt'x @Cp؋“ hcK֮%6؆Pi lZ>A}÷C Ttfo3'MfQ1©zM*~U#NTg69l8sXrQB4:o`!)U?nj|QI~{]UV%$+fA#Uѳ.Vyzu6˽М{Ŭ)Sw7+3,ąu"aih:>=Sꢂm!R(JAb)N.0Z\PN$% fL [k٨1hif9x(p=kέ8*:9fF\}@]RhQ,:E] 1|>#kCnOK(oh " Z"ka[0n޷]䗖{cU;cXK4y&NA' ^ `{:w<@v PNRV%w$v*-;ӣiiih`LJ;>!Gۅ~Jy۾nJO%Zz~wrHDP) S78MC{y*%ur{ !V/(3-E~ ԧTsp~saoÆR~x{=.tz䫢Ͳ;L߰Sܺzܲ/۟i_RXf ;y+fK;W4 I;)UGfC2_ jXhļaWL &+Δd+T=!7:iCne'vh$.X$3~BzN ̉o f,\gc|^Y]$/i9SV%2 ^W~ˑR |[l3̓/2m1 Ao9QkJqDJsZdȪfqf_!==b'JcK {`"F (V[( Dx6skL*+IMc|L4a*{n+jk楨H/ms >W"/ LBζƼwZ Dx^9vYAkGi,MKzs.!9|GT͍ۣfiߒJ6|{9r. :ھ\J.Yw@'O^{.tG3һ5klVFYR& n\BZWoYCv{q1~Njq*rQΧW>۞,c rayauv.;et:9>.9'N:mNdWlIWW_ v{.\q[X첲x?dDg?k疼ڦ}zs빟e_M{DoGyb49eӱZb>K ;߿J^2GSYPzIm`guv$`4pIh|st!c{=G ׍թf!*s(yYB$HKj,`M9$fBLC̣wg.Y "9&|}̗"i a7͆+tq 2;e d,\tQ`8 _\4-R#+ʼnݺP7i,eLhB0#T͠\: 04co ءSu}Ā"'&u.LxO(QݬOM10C~ 8 ~1b*|.=Z&4 d4,'"~2*bJ_ ޠ'22wݤPU1/}dH*da.f͐W4]Sm2P#QE-^Q2dy)vi[ qtx2[ͻsh @:lYR.]]m2 >rE3d\F۝13廍hL型@sfƏm Nmy䞨iLeP f]\D`%GP9tsIrݝ53g4sS<A1J׷V.4wp2-#&UJ g}:;FN%irBLr誊UCv؟erIBG|MLHWFq}X/-yJ UT,VTTXt*_a@30g^3Z59{C;*>" $9 O{o.6tIsV:"U[aXw6ä:U2T<Ф5zߣ$^f.bR䀡+QkmfA*X Fs|{|.Y8{b~:;DwASdz>uIjSu`2avtYKZUtVCLU;nic8YD>my*jpcg){`вmѯLC]m~˛i,OE6u{?E{ \w, 4JcW6Lqm8=N12 ABp>Qz<$zK-dr" CWa( AGI 5+D p=z;> E,mR+;*ҸQ/zi⚈X$ MBc*#>n&bCa{iL$OMrY(z'ya8#lS6BdQ7eӯY<T q.룏/mF8 [.x;<-NPԚZNc٧FAf#4'0-Fߌ4 '.CD^JvRަ0R:vϊ|Wi[\Hi3EdgpMplC{5Th }~q:#@S(Yީe&ªvK^ &@REUOeh5$]u[SȷtHMV>4ι l'e:TB#Z݀ӕ%"تXhpB$+m@M>'Xx-) 9Rp8懵Ů ޶$>) Q4Ph*H,ۅ1%:ȼ)޲pפ/=(@q/C=  4$V!w['b+,M|:4y!j[Puw$Iv +R+%OIs3Flr/vjAdBP ^@za&W+xi-X@WƄADe^!{DУ)>O fzqsf\HPzҀ!lDXx ]nD zNkY#ȍs1y)GHrqp}헄C!x.d)DEig$SӧOD"D5h$=HC>kJB r\佯8 ?rc0I1BBa#1Q)gD (,ƥFx@>ޜTUrhA@W|xjLI= UMZ;-Y $7 ܫc8ǘPrZSM%YݗZ V-s7H #H3v&]w1"][\JCH HVGNFȓ.q$I2x脉iX$i+ r1t]"D(`OD_xgZ3h#Q"47!{CGujzA-ք >AkAq,$vNt:$}>A䠼- IZQc]g|UW W3FK# &g5,:ԾJ_x=)b˂ tn.~Imgf jDŽ͉Rv7pou)x!`Ewx)E/c zW WU5ЭgsmWi(Ln+L9Wg֪{ɂBnĥ"WCfmre5>Dɧ=^U0\sdnB30Q*},%{qݧQdРsO>Pyrdo5Ȱ|o a뾌|l.̗Me 5OmhH2PB6lz`/eMvև0tЅ3#AS4[HuB.&#wpd9CCYc)j7Z7@(aktVgaOVܵ%{k.j g3Z~{Fy1d?#3NA/溪ooďΞ=}g; ΋E[L _o80Ť,h3ƒ8nG'IO|ҁ?leb~dv8/VVqAȕOm|<OfqX(*930`^*)qdl޿>M2txnůC0&f7/ThUhUhUhPUPKE[7DSL>HG co6v2Б흢i|Ef T%㱭P52BOclڛQp/~:g|">4ZQp~vh-1]ʩh.yqM z+ vv|>MC@"|',6uxMdw~hr=nz~}yr5X9x?pW_tܒD[ŹwMdt,HJ吏^#WԗPt O;7b'i,^io4,tՎ2bXl+2+\77OdT[+m4 28ˑ.9J9?55~cyt}8В<-LSCzjc4?$2)oюYo<9X^yjGqV # y*FXehWmuЖǰnm9:mu{ynk>jZ.4䅫NrI.5H6"}Mb5f}Xʱ{ I^wCArK RZn RKM<(v" "|sƲGeEym=UFI+Wr|G (jj!x",h reB/_wl>p}psd뢭dV=Nw;b%6hVMEȚ,dMY>ʐn2 Y˷as]~C 4cJ~-&ooZ1*3`ץ$J ݔR 3c'&I@'gyaD(ooI*%ǸP;|Y!:!8R!<ʥ)$E\>公-c4 ޚ> W۞7_>O෿wjy?W}.yFfNղ~ԆrТr_spOEB(޼WtStz|4:o~]PqKΊ6 A6D_r-Js)6EuP"DK9b:a>TPc;Xvup(qrco.a~jRQ'' Ŭ4٧ ݘ .HJ,&R7X5eIu~Z-U<" ţ7>]X4מ1L]5{i4ے`Փ@5Vd4a7PeW$I~yGBO c)-S 0H-$")t@Asŷj>yΦE,T[<[}t6hi,q@YaF.UeT.;5*Z4 bRGG'Z";;%'&W=~yܵ\KAPBr짋Rc?~iVWz?Ϊp*U.y_-'b^Ѭ/ Jtݕ@T;)'[/FDSN,rwP؄5+c|4L[- )P˲`uwUֶ&V)VJlșLP-Q9: Ô|nSq J;kRߒV,2`'鲵/%/1M"T6h|AWH2rl|8{;'qG~۽_ %yTbd.Sfط:К}=>2i& xQ_0,qy5por}}kDQls]^CqwYTՂ@xͶќ@ ٲ G|.g\SS0. TȘ=2TG3NK&( x(qmKj$Xe!4*Tܰ8h dkI3JAgg$.*FISddL:&9jg荸Zpطa=+]:d~6Z A>jeս0_Ğo*i?_<5kѿ_{|}!VMo&oǧyF*+逿7hQ&R 6Mͽ\yEU2. 8W ('kۨe7^R|f\娵 t8TH~yn s.Ps-oKr!cAS F݈$$][o7+^RE = >`(Nr~jLKd/#YqQUHbceTxU,bȂ@3b( rI U\i#_T)JkLM;2Ʌ.g $n`Amr/w 2+uep68ז* W!P)1\I!bcw`Di|Ty"F?Q߃|M=2bd":̈`͎9%̡a,Շx[a(%x&Fؖ ՇXQC32#!nc A6ׯۋ (q'҈4`_WNՁ?RaO=6[];j˺e3D{{DLR:L7^m*L)8Y' a'lb:{?q۲&PݶIzeT-R;_m\G+ZA,F'/L"м=@Nc\@ڵ_j5(I\LL"f) 3[$\f" |tC=j&<˃XSŜO5FdLDrzW+ O5G9>H|HN~ߛ'a6J-M;"yCu;buƾH4)}GN_Ci+nmHOb_!$"tvzGv[))S6e4[yqvBBq-){;Vw9[!IS,6.v[qgv@Bqݒ)J)Z-؜_sc\]\]~Zύnvχ=/ |`;9s&NgBo{b{o"Bu{_Uq-2 T-!-HU"|}EZKhkn2n`PZ;d^K'Uei4>c0 t]VJ9 ]ץYpDy$/ nܺDTTK&̶= .JuH)/9{8дw@m;'Γs;rMʓ *?@yB+",TBzs񾍪Lzx,n)j%9y\O@VA cs^c& Ԍ u`0~u>R |aM&R~ PʃjQ E 12]r< Irh¶VSW`* cDKmvۍzl`L1PCm4ơVkwA$WА獐1ZE4B I61jQM?BUo7jVB+IQ-) +0T@ $Kĉ6Sl4_{o c)-cruǓS߸fqPŞ5n.z!4g @r7.ٞ.o۸StUZ( 8k6e{ט?}}l|6=li^Fft1իS&.K4ÃY)CDs9:%sanuOa 4WsSαROaK #X'CN]3q[+aWn6sFٍ/Ѣ}{62|]<0^IJ?—^@*u/s|r3!h"F@ظEnCϿ{(qƐ3`Ϯi[&]ל\l~{I*YaВ]~a tjܑXl |wrm5Nl#f\p/یx<|f/85 +d՜ ţvq1ڮFhbfϫ7l;(jإ8xd]v3bIelR 0i` R˹r cA45PiaHf67oȣuyZwW͆@ n}J!Xӿ1G)+Y7FUM}Nhl>-C%sV$b+4A΃-ijO,`|dyy3xS 9'3b|=]ҰvzGKs}vUʼT+>RE(rmt}ij0~X[)~Co睄n9%}[IN=gl = kbKȈh򇃬-zQ(%IZpT 8.irʈ}u:jpd1B8LȖY깳HBdë$Fu%KIEOyx*?(,U_Fc_H>KXf^a@9U/DsyoI#ܛtLY}:!UF"{r΢u -(2@b! O5Je!P|~,{xYݓ=KbNT!VF,qsކiXC1wD0N4Hrw`ԉ0=Hqk5T#3)ϥӱLdԮ56 +G{O՘2/`5%FaْPx!QxM`Ad{P΋Ʌ e`F V5P%0hxExCp~ 0 ].)6xQ}q.ǿw7[VJPDpX! Y韧gM`B-X=?O|w#߬NTA0*tp)aQc0֟꯷hIZJR`x[>PG`ps~Ɗ3^֤f{<͖R(}+A;(a[$O$9!i]F>q-:a3|xOά]Jmů矾T~SGasr}-7G? K=>6ZŖ7Wf6XUm|Z.ߜp,o&s[֧]=|bmAtcᕛhKkqE~om~]W_z{|XsIꩲ 1 fFjkOֆnM#Cs}yif8s<kzBiӄSys8a&nIi[SChLs4"+9DxSfS8@ȵ >]g<-Rݦ$=>gl^cNj}İO%ג`[N7ܫ&VmwխBTlx \Mξ_Ŷ'lv<9)0 SOaKZhs;%}09`aùdB=?J& a1s]Ceuw`XXW@_ѥ VxoOW^dXuLW4ӣX.%LbTw}g"N]oQET_UX6o ZoZ_=_T"zM\S^ډ<Ģ^yopTbMJW-hقڊ]`9%2$]8Uk%cBZ$t HyH3VR'&$aTL1`$ocv ņ ɊWUe]I3Tj [&+#w`La;VAZĻZeX Qj`GBkOFjǚ)bj"Znp#.9׊&+ |%"6!aM02C1<ڊB8]~.$$ K-q%;㙥 ~*ӆ 10940ms (15:22:00.891) Feb 18 15:22:00 crc kubenswrapper[4968]: Trace[48693357]: [10.940879619s] [10.940879619s] END Feb 18 15:22:00 crc kubenswrapper[4968]: I0218 15:22:00.891637 4968 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 18 15:22:00 crc kubenswrapper[4968]: I0218 15:22:00.892424 4968 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 18 15:22:00 crc kubenswrapper[4968]: I0218 15:22:00.895666 4968 trace.go:236] Trace[293836069]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Feb-2026 15:21:50.042) (total time: 10853ms): Feb 18 15:22:00 crc kubenswrapper[4968]: Trace[293836069]: ---"Objects listed" error: 10853ms (15:22:00.895) Feb 18 15:22:00 crc kubenswrapper[4968]: Trace[293836069]: [10.853157516s] [10.853157516s] END Feb 18 15:22:00 crc kubenswrapper[4968]: I0218 15:22:00.895698 4968 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 18 15:22:00 crc kubenswrapper[4968]: E0218 15:22:00.898600 4968 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 18 15:22:00 crc kubenswrapper[4968]: I0218 15:22:00.900533 4968 trace.go:236] Trace[798252892]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Feb-2026 15:21:50.140) (total time: 10759ms): Feb 18 15:22:00 crc kubenswrapper[4968]: Trace[798252892]: ---"Objects listed" error: 10759ms (15:22:00.900) Feb 18 15:22:00 crc kubenswrapper[4968]: Trace[798252892]: [10.759899435s] [10.759899435s] END Feb 18 15:22:00 crc kubenswrapper[4968]: I0218 15:22:00.900558 4968 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 18 15:22:00 crc kubenswrapper[4968]: I0218 15:22:00.963337 4968 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 18 15:22:00 crc kubenswrapper[4968]: I0218 15:22:00.982150 4968 csr.go:261] certificate signing request csr-qr2rn is approved, waiting to be issued Feb 18 15:22:00 crc kubenswrapper[4968]: I0218 15:22:00.991417 4968 csr.go:257] certificate signing request csr-qr2rn is issued Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.016858 4968 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 18 15:22:01 crc kubenswrapper[4968]: W0218 15:22:01.017096 4968 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.CSIDriver ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 18 15:22:01 crc kubenswrapper[4968]: W0218 15:22:01.017134 4968 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 18 15:22:01 crc kubenswrapper[4968]: W0218 15:22:01.017169 4968 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Node ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.017256 4968 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events/crc.18956078fd2dc7d3\": read tcp 38.129.56.151:48508->38.129.56.151:6443: use of closed network connection" event="&Event{ObjectMeta:{crc.18956078fd2dc7d3 default 26194 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-18 15:21:41 +0000 UTC,LastTimestamp:2026-02-18 15:21:41.335882426 +0000 UTC m=+0.721327278,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 18 15:22:01 crc kubenswrapper[4968]: W0218 15:22:01.017394 4968 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Service ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.144583 4968 apiserver.go:52] "Watching apiserver" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.155903 4968 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.159050 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.159685 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.159743 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.159819 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.160182 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.160599 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.160644 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.160712 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.160738 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.160823 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.161510 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 10:28:31.437763623 +0000 UTC Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.162452 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.162966 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.164428 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.167488 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.170467 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.173501 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.178670 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.179212 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.179505 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.240905 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-xnhwb"] Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.241221 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-t5rmj"] Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.241351 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.241425 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-t5rmj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.243978 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.244273 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.244421 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.244550 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.244675 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.244972 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.247976 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.249358 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.253296 4968 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.271884 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.290763 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.301051 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304201 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304274 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304373 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304415 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304451 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304478 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304510 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304543 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304577 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304615 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304648 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304680 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304717 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304775 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304814 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304847 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304883 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304920 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304955 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.304985 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305022 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305067 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305102 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305143 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305185 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305231 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305262 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305293 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305324 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305353 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305386 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305452 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305498 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305531 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305565 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305593 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305620 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305643 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305689 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305717 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305809 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305823 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305895 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305926 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305958 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305992 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306021 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306051 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306084 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306118 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306478 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306531 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306565 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306592 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306630 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306665 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306707 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306770 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306808 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306844 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306870 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306899 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306929 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306956 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306996 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307030 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307060 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307094 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307130 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307166 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307195 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307238 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307271 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307305 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307349 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307383 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307417 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307446 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307477 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307504 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307534 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307569 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307601 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307632 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307662 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307702 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307729 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307786 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307817 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307843 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307871 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307902 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307934 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307978 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308014 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308045 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308080 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308109 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308144 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308180 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308212 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308240 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308267 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308292 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308320 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308353 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308379 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308412 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308457 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308501 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308535 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308572 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308604 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308646 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308686 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308730 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308809 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308858 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308895 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308926 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308953 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308983 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309012 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.305839 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306043 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306182 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.309274 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:22:01.809027877 +0000 UTC m=+21.194472739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309325 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309380 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309424 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309496 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309530 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309564 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309603 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309634 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309679 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309709 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309737 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309781 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309811 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309837 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309869 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309900 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309928 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309960 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.309990 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310039 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310074 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310107 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310134 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310164 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310194 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310221 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310254 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310283 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310316 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310344 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310375 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310405 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310435 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310470 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310501 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310529 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310560 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310589 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310627 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310657 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310688 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310720 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.310772 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.311063 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.317907 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318418 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318469 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318505 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318539 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318571 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.320908 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.325700 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.325780 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.325812 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.325862 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.325884 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.325957 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.325981 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326005 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326025 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326045 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326063 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326081 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326098 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326120 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326139 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326180 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326257 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326278 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326296 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326327 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326345 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326363 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326379 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326397 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326414 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326478 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327161 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1390179a-7a57-4696-ab4c-a0c91eeabea2-hosts-file\") pod \"node-resolver-t5rmj\" (UID: \"1390179a-7a57-4696-ab4c-a0c91eeabea2\") " pod="openshift-dns/node-resolver-t5rmj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327256 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327290 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327313 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327335 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327356 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f9bae90c-908f-40fd-8373-4bf7f9aaede6-rootfs\") pod \"machine-config-daemon-xnhwb\" (UID: \"f9bae90c-908f-40fd-8373-4bf7f9aaede6\") " pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327378 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327400 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327422 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrwt5\" (UniqueName: \"kubernetes.io/projected/f9bae90c-908f-40fd-8373-4bf7f9aaede6-kube-api-access-qrwt5\") pod \"machine-config-daemon-xnhwb\" (UID: \"f9bae90c-908f-40fd-8373-4bf7f9aaede6\") " pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327476 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327498 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327518 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f9bae90c-908f-40fd-8373-4bf7f9aaede6-proxy-tls\") pod \"machine-config-daemon-xnhwb\" (UID: \"f9bae90c-908f-40fd-8373-4bf7f9aaede6\") " pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327542 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327560 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327583 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhzqr\" (UniqueName: \"kubernetes.io/projected/1390179a-7a57-4696-ab4c-a0c91eeabea2-kube-api-access-jhzqr\") pod \"node-resolver-t5rmj\" (UID: \"1390179a-7a57-4696-ab4c-a0c91eeabea2\") " pod="openshift-dns/node-resolver-t5rmj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327600 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f9bae90c-908f-40fd-8373-4bf7f9aaede6-mcd-auth-proxy-config\") pod \"machine-config-daemon-xnhwb\" (UID: \"f9bae90c-908f-40fd-8373-4bf7f9aaede6\") " pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327621 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327640 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327660 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327719 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327731 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327743 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327786 4968 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327797 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.327809 4968 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.328589 4968 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.340480 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306238 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306333 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306601 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306693 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306662 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.306707 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307000 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307019 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307142 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307294 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307319 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307475 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307512 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307583 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.347238 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307701 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308048 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308051 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308041 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.307734 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308143 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308322 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308354 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308411 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308463 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308705 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.308922 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.316992 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.317159 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.317181 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.317232 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.317393 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.317404 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.317623 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318091 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318247 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318249 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318353 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318353 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.317902 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318375 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318538 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.320563 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.320693 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.320696 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.320813 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.320841 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.320862 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.321040 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.321063 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.321257 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.321338 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.321648 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.321712 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.321763 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.322111 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.348132 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.322289 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.322321 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.322355 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.322593 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.322881 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.322910 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.323047 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.318602 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.323320 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.323475 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.323706 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.323867 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.324112 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.324133 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.324142 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.324558 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.324615 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.325657 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326141 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326201 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326449 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326456 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326535 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.326941 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.328215 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.328425 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.328495 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.328640 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.328833 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.329143 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.329179 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.329209 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.329237 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.329281 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.330285 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.330288 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.330539 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.330613 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.330643 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.330868 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.331217 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.331529 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.332017 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.332129 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.332543 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.332734 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.332913 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.333179 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.333244 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.333267 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.333282 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.333676 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.333767 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.333952 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.335024 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.335858 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.335988 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.336247 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.336385 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.336630 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.336685 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.336841 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.337072 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.337168 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.337732 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.338605 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.338646 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.338816 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.339948 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.340082 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.340388 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.340928 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.343410 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.344309 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.344555 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.344719 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.344763 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.345018 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.345476 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.346081 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.346116 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.346199 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.346360 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.346565 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.347381 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.347675 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.347848 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.347863 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.348308 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.348391 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.348612 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.348587 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.348865 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.349697 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.350494 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.350679 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.350794 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.350823 4968 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.350878 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:01.850858656 +0000 UTC m=+21.236303518 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.351009 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.351122 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.351206 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.351509 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.351527 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.351703 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.351826 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.351938 4968 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.351972 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:01.851963228 +0000 UTC m=+21.237408090 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.351992 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.352072 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.352353 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.352404 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.352596 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.347909 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.353012 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.353033 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.353163 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.353170 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.353374 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.353487 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.353519 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.353685 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.353077 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.354464 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.356342 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.356585 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.361016 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.356597 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.364365 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.364855 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.364885 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.364951 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.365140 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.365164 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.365491 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.365534 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.366192 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.366207 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.366219 4968 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.366266 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:01.866251864 +0000 UTC m=+21.251696726 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.367163 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.368295 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.368475 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.368534 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.370275 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.376220 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.378862 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.378895 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.378983 4968 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.379071 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:01.879044856 +0000 UTC m=+21.264489718 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.379924 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.381840 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.382330 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.390257 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.395237 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.409585 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.417722 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.428803 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.428837 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f9bae90c-908f-40fd-8373-4bf7f9aaede6-rootfs\") pod \"machine-config-daemon-xnhwb\" (UID: \"f9bae90c-908f-40fd-8373-4bf7f9aaede6\") " pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.428863 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrwt5\" (UniqueName: \"kubernetes.io/projected/f9bae90c-908f-40fd-8373-4bf7f9aaede6-kube-api-access-qrwt5\") pod \"machine-config-daemon-xnhwb\" (UID: \"f9bae90c-908f-40fd-8373-4bf7f9aaede6\") " pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.428878 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f9bae90c-908f-40fd-8373-4bf7f9aaede6-proxy-tls\") pod \"machine-config-daemon-xnhwb\" (UID: \"f9bae90c-908f-40fd-8373-4bf7f9aaede6\") " pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.428895 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.428912 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhzqr\" (UniqueName: \"kubernetes.io/projected/1390179a-7a57-4696-ab4c-a0c91eeabea2-kube-api-access-jhzqr\") pod \"node-resolver-t5rmj\" (UID: \"1390179a-7a57-4696-ab4c-a0c91eeabea2\") " pod="openshift-dns/node-resolver-t5rmj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.428922 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f9bae90c-908f-40fd-8373-4bf7f9aaede6-rootfs\") pod \"machine-config-daemon-xnhwb\" (UID: \"f9bae90c-908f-40fd-8373-4bf7f9aaede6\") " pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.428937 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f9bae90c-908f-40fd-8373-4bf7f9aaede6-mcd-auth-proxy-config\") pod \"machine-config-daemon-xnhwb\" (UID: \"f9bae90c-908f-40fd-8373-4bf7f9aaede6\") " pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.428924 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.428967 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1390179a-7a57-4696-ab4c-a0c91eeabea2-hosts-file\") pod \"node-resolver-t5rmj\" (UID: \"1390179a-7a57-4696-ab4c-a0c91eeabea2\") " pod="openshift-dns/node-resolver-t5rmj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.428997 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429039 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1390179a-7a57-4696-ab4c-a0c91eeabea2-hosts-file\") pod \"node-resolver-t5rmj\" (UID: \"1390179a-7a57-4696-ab4c-a0c91eeabea2\") " pod="openshift-dns/node-resolver-t5rmj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429007 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429078 4968 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429095 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429110 4968 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429125 4968 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429138 4968 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429150 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429167 4968 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429180 4968 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429192 4968 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429205 4968 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429217 4968 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429263 4968 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429277 4968 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429289 4968 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429303 4968 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429315 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429327 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429338 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429349 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429360 4968 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429373 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429391 4968 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429402 4968 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429414 4968 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429426 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429437 4968 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429448 4968 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429458 4968 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429470 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429486 4968 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429499 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429512 4968 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429524 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429535 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429548 4968 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429560 4968 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429571 4968 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429581 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429592 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429605 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429616 4968 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429627 4968 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429637 4968 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429648 4968 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429659 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429671 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429683 4968 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429687 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f9bae90c-908f-40fd-8373-4bf7f9aaede6-mcd-auth-proxy-config\") pod \"machine-config-daemon-xnhwb\" (UID: \"f9bae90c-908f-40fd-8373-4bf7f9aaede6\") " pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429696 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429712 4968 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429724 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429735 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429777 4968 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429790 4968 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429802 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429813 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429826 4968 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429839 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429851 4968 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429900 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429915 4968 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429927 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429937 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429948 4968 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429958 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429969 4968 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429981 4968 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.429993 4968 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430006 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430017 4968 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430028 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430040 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430052 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430065 4968 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430077 4968 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430088 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430099 4968 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430112 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430123 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430136 4968 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430128 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430150 4968 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430252 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430289 4968 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430302 4968 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430316 4968 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430328 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430340 4968 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430352 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430364 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430377 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430391 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430403 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430415 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430426 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430437 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430449 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430460 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430472 4968 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430484 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430497 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430509 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430522 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430534 4968 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430545 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430557 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430568 4968 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430579 4968 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430590 4968 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430603 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430614 4968 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430626 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430638 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430650 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430661 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430672 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430684 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430695 4968 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430708 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430719 4968 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430731 4968 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430763 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430776 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430787 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430799 4968 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430813 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430825 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430836 4968 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430847 4968 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430859 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430869 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430880 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430891 4968 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430903 4968 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430914 4968 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430925 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430937 4968 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430947 4968 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430957 4968 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430969 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430979 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.430990 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431003 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431014 4968 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431025 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431036 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431790 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431806 4968 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431818 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431830 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431842 4968 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431863 4968 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431874 4968 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431885 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431903 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431941 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431957 4968 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431970 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431982 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.431997 4968 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432007 4968 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432019 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432031 4968 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432043 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432054 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432065 4968 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432078 4968 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432092 4968 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432107 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432118 4968 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432131 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432145 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432161 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432172 4968 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432186 4968 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432198 4968 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432210 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432222 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432233 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432246 4968 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432258 4968 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432270 4968 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432282 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432293 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432306 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432318 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432328 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432339 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.432350 4968 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.433224 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f9bae90c-908f-40fd-8373-4bf7f9aaede6-proxy-tls\") pod \"machine-config-daemon-xnhwb\" (UID: \"f9bae90c-908f-40fd-8373-4bf7f9aaede6\") " pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.441548 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.446037 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhzqr\" (UniqueName: \"kubernetes.io/projected/1390179a-7a57-4696-ab4c-a0c91eeabea2-kube-api-access-jhzqr\") pod \"node-resolver-t5rmj\" (UID: \"1390179a-7a57-4696-ab4c-a0c91eeabea2\") " pod="openshift-dns/node-resolver-t5rmj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.446205 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrwt5\" (UniqueName: \"kubernetes.io/projected/f9bae90c-908f-40fd-8373-4bf7f9aaede6-kube-api-access-qrwt5\") pod \"machine-config-daemon-xnhwb\" (UID: \"f9bae90c-908f-40fd-8373-4bf7f9aaede6\") " pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.451606 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.461701 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.473011 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.478855 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.479045 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.482847 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.501410 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.511206 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.515489 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.527942 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 15:22:01 crc kubenswrapper[4968]: W0218 15:22:01.529796 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-2feb4df6c711c1e29018530ca2aebdc8df25029a794c3cf9cf7a92ff6fe37961 WatchSource:0}: Error finding container 2feb4df6c711c1e29018530ca2aebdc8df25029a794c3cf9cf7a92ff6fe37961: Status 404 returned error can't find the container with id 2feb4df6c711c1e29018530ca2aebdc8df25029a794c3cf9cf7a92ff6fe37961 Feb 18 15:22:01 crc kubenswrapper[4968]: W0218 15:22:01.547549 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-07f078aa57b4abf3e29b2309b6ff6401cc939c01d8a733f3cfa73d9000b04b54 WatchSource:0}: Error finding container 07f078aa57b4abf3e29b2309b6ff6401cc939c01d8a733f3cfa73d9000b04b54: Status 404 returned error can't find the container with id 07f078aa57b4abf3e29b2309b6ff6401cc939c01d8a733f3cfa73d9000b04b54 Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.554696 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.560123 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-t5rmj" Feb 18 15:22:01 crc kubenswrapper[4968]: W0218 15:22:01.590387 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1390179a_7a57_4696_ab4c_a0c91eeabea2.slice/crio-82eea7d89dd728af48b772c51e1b80e8216da69d4e3a8773b0daee6c1e5b2edc WatchSource:0}: Error finding container 82eea7d89dd728af48b772c51e1b80e8216da69d4e3a8773b0daee6c1e5b2edc: Status 404 returned error can't find the container with id 82eea7d89dd728af48b772c51e1b80e8216da69d4e3a8773b0daee6c1e5b2edc Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.623987 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-z2jkk"] Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.627253 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-5rzpj"] Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.627442 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.629712 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.629855 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.629921 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.629974 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.630073 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.630158 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.630353 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-m2qq8"] Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.630587 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.630646 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.632297 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.633018 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.633301 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.637735 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.638310 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.639015 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.639244 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.639460 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.646387 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.671371 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.685191 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.700977 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.711697 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.722882 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.735430 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738391 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-systemd-units\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738433 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-etc-openvswitch\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738481 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6694896c-93a1-47ac-a079-20501cf9909e-cnibin\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738549 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-var-lib-cni-multus\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738583 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovnkube-script-lib\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738629 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6694896c-93a1-47ac-a079-20501cf9909e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738666 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-run-netns\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738706 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6694896c-93a1-47ac-a079-20501cf9909e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738727 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59cjh\" (UniqueName: \"kubernetes.io/projected/6694896c-93a1-47ac-a079-20501cf9909e-kube-api-access-59cjh\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738769 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-openvswitch\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738791 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-kubelet\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738819 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-run-ovn-kubernetes\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738867 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-cni-bin\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738941 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.738967 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-run-multus-certs\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.739018 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6694896c-93a1-47ac-a079-20501cf9909e-os-release\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.739049 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-cni-netd\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.739094 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-multus-cni-dir\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.739124 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-var-lib-cni-bin\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.739170 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6694896c-93a1-47ac-a079-20501cf9909e-system-cni-dir\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.741724 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-node-log\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.741826 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovnkube-config\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.741847 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-var-lib-kubelet\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.741865 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2d01781a-6a78-49a2-80c7-9ac02c810e3f-multus-daemon-config\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.741881 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6694896c-93a1-47ac-a079-20501cf9909e-cni-binary-copy\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.741936 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-system-cni-dir\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.741973 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2d01781a-6a78-49a2-80c7-9ac02c810e3f-cni-binary-copy\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.741995 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-run-netns\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742022 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-multus-socket-dir-parent\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742054 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-run-k8s-cni-cncf-io\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742087 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsdhc\" (UniqueName: \"kubernetes.io/projected/2d01781a-6a78-49a2-80c7-9ac02c810e3f-kube-api-access-xsdhc\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742115 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-slash\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742142 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-log-socket\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742216 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hmgv\" (UniqueName: \"kubernetes.io/projected/435c6f94-b91e-4ce0-8407-5227f3a5078f-kube-api-access-5hmgv\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742272 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-cnibin\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742294 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-systemd\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742313 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-var-lib-openvswitch\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742338 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-os-release\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742376 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-ovn\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742397 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-multus-conf-dir\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742412 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-etc-kubernetes\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742438 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-hostroot\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742482 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-env-overrides\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.742497 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovn-node-metrics-cert\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.746734 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.755934 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.766025 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.777536 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.790001 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.802819 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.816913 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.829386 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.840350 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843265 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843370 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-etc-kubernetes\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843395 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-ovn\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843418 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-multus-conf-dir\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843435 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovn-node-metrics-cert\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843451 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-ovn\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843463 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-etc-kubernetes\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.843479 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:22:02.84343621 +0000 UTC m=+22.228881082 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843506 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-multus-conf-dir\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843559 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-hostroot\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843594 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-hostroot\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843630 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-env-overrides\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843668 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-systemd-units\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843693 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-etc-openvswitch\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843724 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-var-lib-cni-multus\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843777 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6694896c-93a1-47ac-a079-20501cf9909e-cnibin\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843834 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovnkube-script-lib\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843861 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6694896c-93a1-47ac-a079-20501cf9909e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843900 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-run-netns\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843919 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6694896c-93a1-47ac-a079-20501cf9909e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843935 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59cjh\" (UniqueName: \"kubernetes.io/projected/6694896c-93a1-47ac-a079-20501cf9909e-kube-api-access-59cjh\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843958 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-openvswitch\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843975 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-run-multus-certs\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.843991 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-kubelet\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844011 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-run-ovn-kubernetes\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844032 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-cni-bin\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844054 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844073 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6694896c-93a1-47ac-a079-20501cf9909e-system-cni-dir\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844092 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6694896c-93a1-47ac-a079-20501cf9909e-os-release\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844111 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-cni-netd\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844128 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-multus-cni-dir\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844146 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-var-lib-cni-bin\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844169 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-node-log\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844188 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovnkube-config\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844207 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-var-lib-kubelet\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844225 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2d01781a-6a78-49a2-80c7-9ac02c810e3f-multus-daemon-config\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844225 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-run-multus-certs\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844239 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6694896c-93a1-47ac-a079-20501cf9909e-cni-binary-copy\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844285 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-run-netns\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844316 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-system-cni-dir\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844343 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-env-overrides\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844344 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2d01781a-6a78-49a2-80c7-9ac02c810e3f-cni-binary-copy\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844412 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-cnibin\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844437 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-multus-socket-dir-parent\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844459 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-run-k8s-cni-cncf-io\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844482 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsdhc\" (UniqueName: \"kubernetes.io/projected/2d01781a-6a78-49a2-80c7-9ac02c810e3f-kube-api-access-xsdhc\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844510 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-slash\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844531 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-log-socket\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844530 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-kubelet\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844556 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hmgv\" (UniqueName: \"kubernetes.io/projected/435c6f94-b91e-4ce0-8407-5227f3a5078f-kube-api-access-5hmgv\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844582 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-systemd\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844588 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-run-ovn-kubernetes\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844601 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-var-lib-openvswitch\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844625 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-os-release\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844634 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-cni-bin\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844700 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-cni-netd\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844838 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-cnibin\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844883 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-multus-cni-dir\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844920 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-run-k8s-cni-cncf-io\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844929 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-multus-socket-dir-parent\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.844965 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-var-lib-cni-bin\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845005 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-node-log\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845101 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6694896c-93a1-47ac-a079-20501cf9909e-os-release\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845168 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6694896c-93a1-47ac-a079-20501cf9909e-cni-binary-copy\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845172 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2d01781a-6a78-49a2-80c7-9ac02c810e3f-cni-binary-copy\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845277 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-var-lib-cni-multus\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845284 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-slash\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845301 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-log-socket\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845310 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6694896c-93a1-47ac-a079-20501cf9909e-cnibin\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845368 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-os-release\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845391 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-systemd\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845414 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-var-lib-kubelet\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845436 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-var-lib-openvswitch\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845463 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6694896c-93a1-47ac-a079-20501cf9909e-system-cni-dir\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845498 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845587 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovnkube-config\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845625 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-host-run-netns\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845644 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-etc-openvswitch\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845804 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-openvswitch\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845811 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2d01781a-6a78-49a2-80c7-9ac02c810e3f-system-cni-dir\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845828 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-run-netns\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845863 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-systemd-units\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.845868 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovnkube-script-lib\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.846136 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6694896c-93a1-47ac-a079-20501cf9909e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.846232 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2d01781a-6a78-49a2-80c7-9ac02c810e3f-multus-daemon-config\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.848858 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovn-node-metrics-cert\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.851407 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.859660 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.860738 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsdhc\" (UniqueName: \"kubernetes.io/projected/2d01781a-6a78-49a2-80c7-9ac02c810e3f-kube-api-access-xsdhc\") pod \"multus-m2qq8\" (UID: \"2d01781a-6a78-49a2-80c7-9ac02c810e3f\") " pod="openshift-multus/multus-m2qq8" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.868991 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6694896c-93a1-47ac-a079-20501cf9909e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.871075 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hmgv\" (UniqueName: \"kubernetes.io/projected/435c6f94-b91e-4ce0-8407-5227f3a5078f-kube-api-access-5hmgv\") pod \"ovnkube-node-z2jkk\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.872581 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59cjh\" (UniqueName: \"kubernetes.io/projected/6694896c-93a1-47ac-a079-20501cf9909e-kube-api-access-59cjh\") pod \"multus-additional-cni-plugins-5rzpj\" (UID: \"6694896c-93a1-47ac-a079-20501cf9909e\") " pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.880438 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.892282 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.909416 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.919109 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.945566 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.945620 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.945640 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.945663 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.945783 4968 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.945804 4968 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.945849 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:02.945834442 +0000 UTC m=+22.331279304 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.945863 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.945887 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:02.945868553 +0000 UTC m=+22.331313415 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.945889 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.945910 4968 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.945965 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:02.945956176 +0000 UTC m=+22.331401038 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.945987 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.946054 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.946099 4968 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:01 crc kubenswrapper[4968]: E0218 15:22:01.946211 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:02.946181312 +0000 UTC m=+22.331626174 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.961838 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.970825 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.992452 4968 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-18 15:17:00 +0000 UTC, rotation deadline is 2026-12-21 08:37:09.897702876 +0000 UTC Feb 18 15:22:01 crc kubenswrapper[4968]: I0218 15:22:01.992513 4968 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7337h15m7.905192264s for next certificate rotation Feb 18 15:22:01 crc kubenswrapper[4968]: W0218 15:22:01.993503 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod435c6f94_b91e_4ce0_8407_5227f3a5078f.slice/crio-a5a166ddc4cb02e8be575a1382f44a5f44030729235c4926c78175404505570f WatchSource:0}: Error finding container a5a166ddc4cb02e8be575a1382f44a5f44030729235c4926c78175404505570f: Status 404 returned error can't find the container with id a5a166ddc4cb02e8be575a1382f44a5f44030729235c4926c78175404505570f Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:01.999391 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-m2qq8" Feb 18 15:22:02 crc kubenswrapper[4968]: W0218 15:22:02.029405 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d01781a_6a78_49a2_80c7_9ac02c810e3f.slice/crio-eca7608d6c41b8c05691e779653741ca257a1a10c5cd394ff72f35d1a3fa36e2 WatchSource:0}: Error finding container eca7608d6c41b8c05691e779653741ca257a1a10c5cd394ff72f35d1a3fa36e2: Status 404 returned error can't find the container with id eca7608d6c41b8c05691e779653741ca257a1a10c5cd394ff72f35d1a3fa36e2 Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.162437 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 05:54:07.683315084 +0000 UTC Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.229934 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.229999 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.230082 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.230199 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.379730 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.379812 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"a5a166ddc4cb02e8be575a1382f44a5f44030729235c4926c78175404505570f"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.384893 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-t5rmj" event={"ID":"1390179a-7a57-4696-ab4c-a0c91eeabea2","Type":"ContainerStarted","Data":"6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.384986 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-t5rmj" event={"ID":"1390179a-7a57-4696-ab4c-a0c91eeabea2","Type":"ContainerStarted","Data":"82eea7d89dd728af48b772c51e1b80e8216da69d4e3a8773b0daee6c1e5b2edc"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.391486 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerStarted","Data":"959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.391564 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerStarted","Data":"940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.391581 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerStarted","Data":"5988c1f53a14ebd77a5c1ccdda32f97d87e20be8e033586248196175285d528e"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.396296 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2feb4df6c711c1e29018530ca2aebdc8df25029a794c3cf9cf7a92ff6fe37961"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.402136 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.402178 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.402191 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"07f078aa57b4abf3e29b2309b6ff6401cc939c01d8a733f3cfa73d9000b04b54"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.407517 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.407581 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1fb765c771d5dd513887f2945ab4453cc00e46edb13665a8f44b7e673927a4cb"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.409318 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2qq8" event={"ID":"2d01781a-6a78-49a2-80c7-9ac02c810e3f","Type":"ContainerStarted","Data":"7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.409367 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2qq8" event={"ID":"2d01781a-6a78-49a2-80c7-9ac02c810e3f","Type":"ContainerStarted","Data":"eca7608d6c41b8c05691e779653741ca257a1a10c5cd394ff72f35d1a3fa36e2"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.410824 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" event={"ID":"6694896c-93a1-47ac-a079-20501cf9909e","Type":"ContainerStarted","Data":"074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.410885 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" event={"ID":"6694896c-93a1-47ac-a079-20501cf9909e","Type":"ContainerStarted","Data":"107d566e462f5d5324ebea19bb75a697121b71e06b02850ae07512061093ae11"} Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.425863 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.458357 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.509812 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.576216 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.602791 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.618116 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.632240 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.644718 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.651382 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.663162 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.674389 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.688232 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.697936 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.855960 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.856189 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:22:04.856156004 +0000 UTC m=+24.241600916 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.957428 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.957470 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.957497 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.957537 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.957645 4968 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.957694 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:04.957679911 +0000 UTC m=+24.343124773 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.958058 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.958075 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.958086 4968 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.958112 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:04.958105373 +0000 UTC m=+24.343550235 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.958153 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.958162 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.958169 4968 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.958189 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:04.958182775 +0000 UTC m=+24.343627627 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.958278 4968 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:02 crc kubenswrapper[4968]: E0218 15:22:02.958374 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:04.95835226 +0000 UTC m=+24.343797292 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.967446 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:02 crc kubenswrapper[4968]: I0218 15:22:02.985945 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.004117 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.030090 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.045373 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.064872 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.080633 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.094480 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.103929 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.113782 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.128268 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.162611 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 22:42:33.936530228 +0000 UTC Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.230258 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:03 crc kubenswrapper[4968]: E0218 15:22:03.230463 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.235131 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.244172 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.245075 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.245879 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.252037 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.252645 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.253298 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.254834 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.255790 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.256324 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.257010 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.257892 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.258648 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.259510 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.260056 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.260587 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.261219 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.261650 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.262225 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.262856 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.282397 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.283345 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.360165 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.361116 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.361884 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.362799 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.363616 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.364173 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.364812 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.365384 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.365977 4968 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.366093 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.389771 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.395589 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.396115 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.418438 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502" exitCode=0 Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.431590 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.443008 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.458134 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.463187 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.464085 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.465089 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.465732 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.466802 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.468495 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.469989 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.482196 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.496405 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.497616 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.498515 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.500090 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.501445 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.502343 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.503325 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.504005 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.504609 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.505169 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.505820 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.506430 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.506998 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.507550 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502"} Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.509995 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.522112 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.530881 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.547478 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.559898 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.584378 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.598560 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.628803 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.645015 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.656869 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.671093 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.683388 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.700038 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.712081 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.726726 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.737265 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.749381 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.760238 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 15:22:03 crc kubenswrapper[4968]: I0218 15:22:03.802669 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:03Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.162976 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 10:37:49.788746668 +0000 UTC Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.230651 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.230689 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.230816 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.231188 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.424636 4968 generic.go:334] "Generic (PLEG): container finished" podID="6694896c-93a1-47ac-a079-20501cf9909e" containerID="074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd" exitCode=0 Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.424694 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" event={"ID":"6694896c-93a1-47ac-a079-20501cf9909e","Type":"ContainerDied","Data":"074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd"} Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.429679 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608"} Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.429729 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c"} Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.429763 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c"} Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.429776 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a"} Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.429789 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2"} Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.429802 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f"} Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.443522 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.476385 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.494532 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.513983 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.529964 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.547462 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.562737 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.579198 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.594353 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.608804 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.626427 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.646912 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.873844 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.876979 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.877221 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:22:08.877182493 +0000 UTC m=+28.262627535 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.886122 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.891613 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.895166 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.913048 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.932546 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.949191 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.966822 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.978692 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.978755 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.978777 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.978798 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.978878 4968 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.978928 4968 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.978942 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:08.978924516 +0000 UTC m=+28.364369378 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.979070 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.979129 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.979136 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.979148 4968 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.979158 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.979094 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:08.97904044 +0000 UTC m=+28.364485452 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.979176 4968 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.979231 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:08.979209615 +0000 UTC m=+28.364654647 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:04 crc kubenswrapper[4968]: E0218 15:22:04.979285 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:08.979271407 +0000 UTC m=+28.364716449 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.982125 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:04 crc kubenswrapper[4968]: I0218 15:22:04.996667 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:04Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.013602 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.031249 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.045191 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.066400 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.084119 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.101221 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.116637 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.130912 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.146176 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.162613 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.168100 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 20:24:32.992468238 +0000 UTC Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.179386 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.204478 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.226030 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.232128 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:05 crc kubenswrapper[4968]: E0218 15:22:05.232221 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.290017 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.356113 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.380088 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.395418 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.415488 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.441075 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" event={"ID":"6694896c-93a1-47ac-a079-20501cf9909e","Type":"ContainerStarted","Data":"8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d"} Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.444108 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf"} Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.463096 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.480532 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.497887 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.514237 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.532579 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.562075 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.576706 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.617933 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.636190 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.654119 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.670397 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.684437 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.705036 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.721363 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.747939 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.761055 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.788811 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.806220 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.854119 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.879936 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.917318 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.962458 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:05 crc kubenswrapper[4968]: I0218 15:22:05.997580 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:05Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.039498 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.077385 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.115716 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.169234 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 04:30:40.07430094 +0000 UTC Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.230222 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.230230 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:06 crc kubenswrapper[4968]: E0218 15:22:06.230445 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:06 crc kubenswrapper[4968]: E0218 15:22:06.230586 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.448183 4968 generic.go:334] "Generic (PLEG): container finished" podID="6694896c-93a1-47ac-a079-20501cf9909e" containerID="8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d" exitCode=0 Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.448279 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" event={"ID":"6694896c-93a1-47ac-a079-20501cf9909e","Type":"ContainerDied","Data":"8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d"} Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.465194 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.482780 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.497971 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.514732 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.529107 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.540902 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.566477 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.583044 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.600130 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.620416 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.632244 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.656032 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:06 crc kubenswrapper[4968]: I0218 15:22:06.670648 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:06Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.169684 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 01:29:12.811223389 +0000 UTC Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.230371 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:07 crc kubenswrapper[4968]: E0218 15:22:07.230523 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.299587 4968 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.301761 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.301808 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.301820 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.301948 4968 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.302302 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.305973 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.313874 4968 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.314142 4968 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.315053 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.315163 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.315195 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.315206 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.315221 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.315234 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:07Z","lastTransitionTime":"2026-02-18T15:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.319277 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: E0218 15:22:07.328077 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.330665 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.342511 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.351414 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.361363 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.361410 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.361422 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.361439 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.361453 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:07Z","lastTransitionTime":"2026-02-18T15:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.364718 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: E0218 15:22:07.374536 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.378256 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.378369 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.378397 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.378454 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.378468 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:07Z","lastTransitionTime":"2026-02-18T15:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.388816 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: E0218 15:22:07.395508 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.398817 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.398871 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.398883 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.398901 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.398935 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:07Z","lastTransitionTime":"2026-02-18T15:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.403839 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: E0218 15:22:07.411823 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.414676 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.414701 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.414708 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.414738 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.414775 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:07Z","lastTransitionTime":"2026-02-18T15:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.421880 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: E0218 15:22:07.426865 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: E0218 15:22:07.427139 4968 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.428609 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.428648 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.428658 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.428672 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.428681 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:07Z","lastTransitionTime":"2026-02-18T15:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.435769 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.448553 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.458519 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49"} Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.462365 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.462433 4968 generic.go:334] "Generic (PLEG): container finished" podID="6694896c-93a1-47ac-a079-20501cf9909e" containerID="8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d" exitCode=0 Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.462521 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" event={"ID":"6694896c-93a1-47ac-a079-20501cf9909e","Type":"ContainerDied","Data":"8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d"} Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.475292 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.491207 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.508508 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.534227 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.534273 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.534285 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.534307 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.534321 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:07Z","lastTransitionTime":"2026-02-18T15:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.534698 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.548910 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.569721 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.583999 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.597669 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.611600 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.626986 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.637191 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.637233 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.637243 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.637260 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.637271 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:07Z","lastTransitionTime":"2026-02-18T15:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.639197 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.651864 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.666585 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.696943 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.737316 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.739905 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.739931 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.739938 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.739951 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.739960 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:07Z","lastTransitionTime":"2026-02-18T15:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.775947 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.842322 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.842360 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.842368 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.842382 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.842394 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:07Z","lastTransitionTime":"2026-02-18T15:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.945556 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.945613 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.945645 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.945663 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:07 crc kubenswrapper[4968]: I0218 15:22:07.945677 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:07Z","lastTransitionTime":"2026-02-18T15:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.048948 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.049023 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.049040 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.049067 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.049085 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:08Z","lastTransitionTime":"2026-02-18T15:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.152702 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.152899 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.152971 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.153009 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.153083 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:08Z","lastTransitionTime":"2026-02-18T15:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.171157 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 00:12:54.316057697 +0000 UTC Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.229785 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.229852 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:08 crc kubenswrapper[4968]: E0218 15:22:08.230021 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:08 crc kubenswrapper[4968]: E0218 15:22:08.230228 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.256315 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.256364 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.256377 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.256398 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.256411 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:08Z","lastTransitionTime":"2026-02-18T15:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.359346 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.359401 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.359471 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.359489 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.359499 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:08Z","lastTransitionTime":"2026-02-18T15:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.461584 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.461628 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.461638 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.461652 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.461663 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:08Z","lastTransitionTime":"2026-02-18T15:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.467578 4968 generic.go:334] "Generic (PLEG): container finished" podID="6694896c-93a1-47ac-a079-20501cf9909e" containerID="00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17" exitCode=0 Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.467606 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" event={"ID":"6694896c-93a1-47ac-a079-20501cf9909e","Type":"ContainerDied","Data":"00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17"} Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.494803 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.507722 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.565361 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.565406 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.565416 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.565431 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.565443 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:08Z","lastTransitionTime":"2026-02-18T15:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.580998 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.581362 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-78lnz"] Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.583643 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-78lnz" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.588523 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.589831 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.590230 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.590440 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.612096 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.660767 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.668074 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.668115 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.668129 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.668145 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.668155 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:08Z","lastTransitionTime":"2026-02-18T15:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.677316 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.695479 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.711659 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.726844 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.737128 4968 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.742620 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.752038 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/358a1641-853a-4075-8f59-de12eb00b601-serviceca\") pod \"node-ca-78lnz\" (UID: \"358a1641-853a-4075-8f59-de12eb00b601\") " pod="openshift-image-registry/node-ca-78lnz" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.752077 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/358a1641-853a-4075-8f59-de12eb00b601-host\") pod \"node-ca-78lnz\" (UID: \"358a1641-853a-4075-8f59-de12eb00b601\") " pod="openshift-image-registry/node-ca-78lnz" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.752100 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4wsz\" (UniqueName: \"kubernetes.io/projected/358a1641-853a-4075-8f59-de12eb00b601-kube-api-access-m4wsz\") pod \"node-ca-78lnz\" (UID: \"358a1641-853a-4075-8f59-de12eb00b601\") " pod="openshift-image-registry/node-ca-78lnz" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.757187 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.770463 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.770497 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.770508 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.770523 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.770535 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:08Z","lastTransitionTime":"2026-02-18T15:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.770940 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.782711 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.799019 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.814415 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.835280 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.847807 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.853085 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/358a1641-853a-4075-8f59-de12eb00b601-serviceca\") pod \"node-ca-78lnz\" (UID: \"358a1641-853a-4075-8f59-de12eb00b601\") " pod="openshift-image-registry/node-ca-78lnz" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.853190 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/358a1641-853a-4075-8f59-de12eb00b601-host\") pod \"node-ca-78lnz\" (UID: \"358a1641-853a-4075-8f59-de12eb00b601\") " pod="openshift-image-registry/node-ca-78lnz" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.853330 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/358a1641-853a-4075-8f59-de12eb00b601-host\") pod \"node-ca-78lnz\" (UID: \"358a1641-853a-4075-8f59-de12eb00b601\") " pod="openshift-image-registry/node-ca-78lnz" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.853381 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4wsz\" (UniqueName: \"kubernetes.io/projected/358a1641-853a-4075-8f59-de12eb00b601-kube-api-access-m4wsz\") pod \"node-ca-78lnz\" (UID: \"358a1641-853a-4075-8f59-de12eb00b601\") " pod="openshift-image-registry/node-ca-78lnz" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.854067 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/358a1641-853a-4075-8f59-de12eb00b601-serviceca\") pod \"node-ca-78lnz\" (UID: \"358a1641-853a-4075-8f59-de12eb00b601\") " pod="openshift-image-registry/node-ca-78lnz" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.868545 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.877700 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4wsz\" (UniqueName: \"kubernetes.io/projected/358a1641-853a-4075-8f59-de12eb00b601-kube-api-access-m4wsz\") pod \"node-ca-78lnz\" (UID: \"358a1641-853a-4075-8f59-de12eb00b601\") " pod="openshift-image-registry/node-ca-78lnz" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.879287 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.879313 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.879321 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.879336 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.879344 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:08Z","lastTransitionTime":"2026-02-18T15:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.894228 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.913813 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.934228 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.948328 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.954120 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-78lnz" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.954545 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:08 crc kubenswrapper[4968]: E0218 15:22:08.954707 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:22:16.954684502 +0000 UTC m=+36.340129364 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.964917 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: W0218 15:22:08.974103 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod358a1641_853a_4075_8f59_de12eb00b601.slice/crio-0204bd69fd558002c2c8702c4ec90acf3aacc1a5d52650bef1da037c39910237 WatchSource:0}: Error finding container 0204bd69fd558002c2c8702c4ec90acf3aacc1a5d52650bef1da037c39910237: Status 404 returned error can't find the container with id 0204bd69fd558002c2c8702c4ec90acf3aacc1a5d52650bef1da037c39910237 Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.982177 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.982203 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.982212 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.982229 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.982240 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:08Z","lastTransitionTime":"2026-02-18T15:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.982287 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:08 crc kubenswrapper[4968]: I0218 15:22:08.995649 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.014090 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.029547 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.044515 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.055707 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.055764 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.055789 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.055823 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.055907 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.055930 4968 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.055944 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.055958 4968 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.055965 4968 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.055907 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.056004 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.056013 4968 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.055982 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:17.055966922 +0000 UTC m=+36.441411784 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.056060 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:17.056043304 +0000 UTC m=+36.441488156 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.056073 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:17.056067815 +0000 UTC m=+36.441512677 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.056089 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:17.056081205 +0000 UTC m=+36.441526067 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.062552 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.079510 4968 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.085211 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.085239 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.085249 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.085266 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.085276 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:09Z","lastTransitionTime":"2026-02-18T15:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.171895 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 23:48:22.915035497 +0000 UTC Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.186857 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.186884 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.186894 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.186909 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.186918 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:09Z","lastTransitionTime":"2026-02-18T15:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.231832 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:09 crc kubenswrapper[4968]: E0218 15:22:09.231979 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.290292 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.290338 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.290347 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.290363 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.290375 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:09Z","lastTransitionTime":"2026-02-18T15:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.395818 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.395880 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.395904 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.395937 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.395959 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:09Z","lastTransitionTime":"2026-02-18T15:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.477266 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" event={"ID":"6694896c-93a1-47ac-a079-20501cf9909e","Type":"ContainerStarted","Data":"6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.481061 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.481677 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.481769 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.482937 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-78lnz" event={"ID":"358a1641-853a-4075-8f59-de12eb00b601","Type":"ContainerStarted","Data":"5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.482965 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-78lnz" event={"ID":"358a1641-853a-4075-8f59-de12eb00b601","Type":"ContainerStarted","Data":"0204bd69fd558002c2c8702c4ec90acf3aacc1a5d52650bef1da037c39910237"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.500726 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.500777 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.500786 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.500834 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.500845 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:09Z","lastTransitionTime":"2026-02-18T15:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.502037 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.516505 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.516574 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.528963 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.550527 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.577052 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.593444 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.603278 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.603321 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.603337 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.603355 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.603371 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:09Z","lastTransitionTime":"2026-02-18T15:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.613775 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.634039 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.647633 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.661154 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.673879 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.684538 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.694847 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.705460 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.705504 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.705514 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.705531 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.705543 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:09Z","lastTransitionTime":"2026-02-18T15:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.708687 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.724351 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.739462 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.761651 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.776670 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.804878 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.807433 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.807485 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.807497 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.807519 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.807533 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:09Z","lastTransitionTime":"2026-02-18T15:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.835982 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.878843 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.909520 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.909563 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.909587 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.909605 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.909616 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:09Z","lastTransitionTime":"2026-02-18T15:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.916413 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:09 crc kubenswrapper[4968]: I0218 15:22:09.957245 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.003271 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.013350 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.013400 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.013412 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.013434 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.013446 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:10Z","lastTransitionTime":"2026-02-18T15:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.036572 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.077408 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.115806 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.115838 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.115846 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.115860 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.115870 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:10Z","lastTransitionTime":"2026-02-18T15:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.117088 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.157299 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.172108 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 01:15:27.481184907 +0000 UTC Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.198239 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.218154 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.218199 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.218212 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.218229 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.218241 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:10Z","lastTransitionTime":"2026-02-18T15:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.229675 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:10 crc kubenswrapper[4968]: E0218 15:22:10.229826 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.229679 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:10 crc kubenswrapper[4968]: E0218 15:22:10.230008 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.236333 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.280244 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.320429 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.320468 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.320481 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.320498 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.320509 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:10Z","lastTransitionTime":"2026-02-18T15:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.423760 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.423808 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.423819 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.423837 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.423848 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:10Z","lastTransitionTime":"2026-02-18T15:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.489261 4968 generic.go:334] "Generic (PLEG): container finished" podID="6694896c-93a1-47ac-a079-20501cf9909e" containerID="6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec" exitCode=0 Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.489437 4968 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.490443 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" event={"ID":"6694896c-93a1-47ac-a079-20501cf9909e","Type":"ContainerDied","Data":"6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec"} Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.509362 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.527546 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.527593 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.527605 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.527626 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.527638 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:10Z","lastTransitionTime":"2026-02-18T15:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.538307 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.554323 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.573983 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.593016 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.609718 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.626168 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.631046 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.631085 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.631101 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.631124 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.631141 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:10Z","lastTransitionTime":"2026-02-18T15:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.643448 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.656704 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.676329 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.718057 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.735047 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.735114 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.735143 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.735178 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.735208 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:10Z","lastTransitionTime":"2026-02-18T15:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.771357 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.772326 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.801664 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.839490 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.839533 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.839544 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.839561 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.839571 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:10Z","lastTransitionTime":"2026-02-18T15:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.846766 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.877370 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.915701 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.941980 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.942321 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.942439 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.942544 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.942646 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:10Z","lastTransitionTime":"2026-02-18T15:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.956688 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:10 crc kubenswrapper[4968]: I0218 15:22:10.997448 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.037134 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.044698 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.044733 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.044741 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.044773 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.044785 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:11Z","lastTransitionTime":"2026-02-18T15:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.068188 4968 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.098102 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.138190 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.147212 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.147244 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.147252 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.147264 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.147273 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:11Z","lastTransitionTime":"2026-02-18T15:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.173177 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 07:42:26.346907912 +0000 UTC Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.178991 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.215826 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.229916 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:11 crc kubenswrapper[4968]: E0218 15:22:11.230030 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.249178 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.249208 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.249217 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.249232 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.249242 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:11Z","lastTransitionTime":"2026-02-18T15:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.266436 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.298305 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.343618 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.351351 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.351406 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.351421 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.351441 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.351453 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:11Z","lastTransitionTime":"2026-02-18T15:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.377893 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.415766 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.454239 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.454271 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.454280 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.454294 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.454304 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:11Z","lastTransitionTime":"2026-02-18T15:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.455678 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.497113 4968 generic.go:334] "Generic (PLEG): container finished" podID="6694896c-93a1-47ac-a079-20501cf9909e" containerID="fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8" exitCode=0 Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.497175 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" event={"ID":"6694896c-93a1-47ac-a079-20501cf9909e","Type":"ContainerDied","Data":"fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8"} Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.497374 4968 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.498205 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.544074 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.562985 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.563030 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.563039 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.563086 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.563098 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:11Z","lastTransitionTime":"2026-02-18T15:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.581227 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.619918 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.657171 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.666075 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.666122 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.666131 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.666147 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.666157 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:11Z","lastTransitionTime":"2026-02-18T15:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.700545 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.741295 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.771440 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.771495 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.771507 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.771525 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.771536 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:11Z","lastTransitionTime":"2026-02-18T15:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.780334 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.818327 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.859108 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.876741 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.876824 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.876841 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.876865 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.876880 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:11Z","lastTransitionTime":"2026-02-18T15:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.898049 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.907267 4968 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.963292 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.980460 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.980493 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.980502 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.980517 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:11 crc kubenswrapper[4968]: I0218 15:22:11.980527 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:11Z","lastTransitionTime":"2026-02-18T15:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.040691 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.063607 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.076381 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.085240 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.085311 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.085320 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.085342 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.085354 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:12Z","lastTransitionTime":"2026-02-18T15:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.124027 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.158191 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.174083 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 18:03:50.787621768 +0000 UTC Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.187934 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.187967 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.187976 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.187991 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.188001 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:12Z","lastTransitionTime":"2026-02-18T15:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.197367 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.230019 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.230186 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:12 crc kubenswrapper[4968]: E0218 15:22:12.230254 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:12 crc kubenswrapper[4968]: E0218 15:22:12.230430 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.237031 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.278637 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.290021 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.290053 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.290062 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.290077 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.290085 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:12Z","lastTransitionTime":"2026-02-18T15:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.318265 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.357089 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.393159 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.393581 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.393683 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.393793 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.393884 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:12Z","lastTransitionTime":"2026-02-18T15:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.399963 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.438826 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.478782 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.496850 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.496893 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.496904 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.496925 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.496937 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:12Z","lastTransitionTime":"2026-02-18T15:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.503842 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" event={"ID":"6694896c-93a1-47ac-a079-20501cf9909e","Type":"ContainerStarted","Data":"24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a"} Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.515692 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.555872 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.599821 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.599860 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.599869 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.599882 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.599892 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:12Z","lastTransitionTime":"2026-02-18T15:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.605400 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.643352 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.691691 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.704727 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.704788 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.704800 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.704822 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.704832 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:12Z","lastTransitionTime":"2026-02-18T15:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.717661 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.762222 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.797523 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.807570 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.807631 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.807645 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.807666 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.807680 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:12Z","lastTransitionTime":"2026-02-18T15:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.843606 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.898092 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.909949 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.910020 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.910031 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.910124 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.910135 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:12Z","lastTransitionTime":"2026-02-18T15:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.921625 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.955929 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:12 crc kubenswrapper[4968]: I0218 15:22:12.999314 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:12Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.012615 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.012670 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.012685 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.012708 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.012723 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:13Z","lastTransitionTime":"2026-02-18T15:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.038729 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.081161 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.115451 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.115512 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.115531 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.115560 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.115579 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:13Z","lastTransitionTime":"2026-02-18T15:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.122552 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.158843 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.175185 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 17:59:42.358536107 +0000 UTC Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.196808 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.217833 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.217884 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.217894 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.217911 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.217922 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:13Z","lastTransitionTime":"2026-02-18T15:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.231045 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:13 crc kubenswrapper[4968]: E0218 15:22:13.231186 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.235509 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.279372 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.320479 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.320517 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.320532 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.320552 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.320565 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:13Z","lastTransitionTime":"2026-02-18T15:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.323976 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.423550 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.423583 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.423593 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.423609 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.423620 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:13Z","lastTransitionTime":"2026-02-18T15:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.509816 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/0.log" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.512321 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491" exitCode=1 Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.512377 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491"} Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.513171 4968 scope.go:117] "RemoveContainer" containerID="ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.526545 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.526624 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.526648 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.526679 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.526701 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:13Z","lastTransitionTime":"2026-02-18T15:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.537069 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.564151 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.578279 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.598651 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"message\\\":\\\"ry.go:160\\\\nI0218 15:22:12.837603 6165 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.837842 6165 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 15:22:12.837960 6165 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.838307 6165 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.838537 6165 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.850893 6165 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0218 15:22:12.850942 6165 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:12.850946 6165 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:12.850970 6165 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:12.850977 6165 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 15:22:12.850983 6165 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:12.863924 6165 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.612572 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.627776 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.629207 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.629255 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.629264 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.629279 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.629293 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:13Z","lastTransitionTime":"2026-02-18T15:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.642403 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.658802 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.680048 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.718198 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.731287 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.731322 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.731333 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.731347 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.731358 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:13Z","lastTransitionTime":"2026-02-18T15:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.761544 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.795832 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.833658 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.833734 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.833762 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.833783 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.833796 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:13Z","lastTransitionTime":"2026-02-18T15:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.839805 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.877525 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.917570 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:13Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.936327 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.936371 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.936386 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.936403 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:13 crc kubenswrapper[4968]: I0218 15:22:13.936415 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:13Z","lastTransitionTime":"2026-02-18T15:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.038825 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.038887 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.038905 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.038934 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.038953 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:14Z","lastTransitionTime":"2026-02-18T15:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.142080 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.142165 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.142177 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.142199 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.142210 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:14Z","lastTransitionTime":"2026-02-18T15:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.176237 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 05:30:33.5781029 +0000 UTC Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.230165 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.230173 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:14 crc kubenswrapper[4968]: E0218 15:22:14.230349 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:14 crc kubenswrapper[4968]: E0218 15:22:14.230534 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.244792 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.244830 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.244840 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.244856 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.244867 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:14Z","lastTransitionTime":"2026-02-18T15:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.348317 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.348367 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.348379 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.348398 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.348409 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:14Z","lastTransitionTime":"2026-02-18T15:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.450650 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.450690 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.450700 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.450714 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.450727 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:14Z","lastTransitionTime":"2026-02-18T15:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.522006 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/0.log" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.524898 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7"} Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.525026 4968 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.542037 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.554224 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9"] Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.554993 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.555580 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.555640 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.555654 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.555674 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.555687 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:14Z","lastTransitionTime":"2026-02-18T15:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.558255 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.558371 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.561839 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.578628 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.593662 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.606530 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.618847 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cd365caa-2db2-407a-b264-d18a0d06fe06-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qsjc9\" (UID: \"cd365caa-2db2-407a-b264-d18a0d06fe06\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.618967 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cd365caa-2db2-407a-b264-d18a0d06fe06-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qsjc9\" (UID: \"cd365caa-2db2-407a-b264-d18a0d06fe06\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.619024 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cd365caa-2db2-407a-b264-d18a0d06fe06-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qsjc9\" (UID: \"cd365caa-2db2-407a-b264-d18a0d06fe06\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.619048 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qd8r\" (UniqueName: \"kubernetes.io/projected/cd365caa-2db2-407a-b264-d18a0d06fe06-kube-api-access-7qd8r\") pod \"ovnkube-control-plane-749d76644c-qsjc9\" (UID: \"cd365caa-2db2-407a-b264-d18a0d06fe06\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.619831 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.633725 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.658274 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.658315 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.658326 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.658341 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.658351 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:14Z","lastTransitionTime":"2026-02-18T15:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.659996 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.674156 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.694394 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"message\\\":\\\"ry.go:160\\\\nI0218 15:22:12.837603 6165 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.837842 6165 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 15:22:12.837960 6165 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.838307 6165 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.838537 6165 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.850893 6165 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0218 15:22:12.850942 6165 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:12.850946 6165 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:12.850970 6165 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:12.850977 6165 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 15:22:12.850983 6165 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:12.863924 6165 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.708262 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.719989 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cd365caa-2db2-407a-b264-d18a0d06fe06-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qsjc9\" (UID: \"cd365caa-2db2-407a-b264-d18a0d06fe06\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.720030 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cd365caa-2db2-407a-b264-d18a0d06fe06-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qsjc9\" (UID: \"cd365caa-2db2-407a-b264-d18a0d06fe06\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.720058 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qd8r\" (UniqueName: \"kubernetes.io/projected/cd365caa-2db2-407a-b264-d18a0d06fe06-kube-api-access-7qd8r\") pod \"ovnkube-control-plane-749d76644c-qsjc9\" (UID: \"cd365caa-2db2-407a-b264-d18a0d06fe06\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.720097 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cd365caa-2db2-407a-b264-d18a0d06fe06-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qsjc9\" (UID: \"cd365caa-2db2-407a-b264-d18a0d06fe06\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.720661 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cd365caa-2db2-407a-b264-d18a0d06fe06-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qsjc9\" (UID: \"cd365caa-2db2-407a-b264-d18a0d06fe06\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.720683 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cd365caa-2db2-407a-b264-d18a0d06fe06-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qsjc9\" (UID: \"cd365caa-2db2-407a-b264-d18a0d06fe06\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.725478 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.727261 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cd365caa-2db2-407a-b264-d18a0d06fe06-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qsjc9\" (UID: \"cd365caa-2db2-407a-b264-d18a0d06fe06\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.738841 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qd8r\" (UniqueName: \"kubernetes.io/projected/cd365caa-2db2-407a-b264-d18a0d06fe06-kube-api-access-7qd8r\") pod \"ovnkube-control-plane-749d76644c-qsjc9\" (UID: \"cd365caa-2db2-407a-b264-d18a0d06fe06\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.742151 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.760848 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.760892 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.760904 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.760920 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.760931 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:14Z","lastTransitionTime":"2026-02-18T15:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.762046 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.776787 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.788735 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.816663 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"message\\\":\\\"ry.go:160\\\\nI0218 15:22:12.837603 6165 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.837842 6165 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 15:22:12.837960 6165 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.838307 6165 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.838537 6165 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.850893 6165 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0218 15:22:12.850942 6165 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:12.850946 6165 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:12.850970 6165 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:12.850977 6165 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 15:22:12.850983 6165 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:12.863924 6165 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.831241 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.864325 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.864395 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.864414 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.864438 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.864454 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:14Z","lastTransitionTime":"2026-02-18T15:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.866894 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.877776 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.883870 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: W0218 15:22:14.893008 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd365caa_2db2_407a_b264_d18a0d06fe06.slice/crio-410d93560360f4446fb9e512c75e7244ab482ff7bf25bc2abc7c65d95302bf1f WatchSource:0}: Error finding container 410d93560360f4446fb9e512c75e7244ab482ff7bf25bc2abc7c65d95302bf1f: Status 404 returned error can't find the container with id 410d93560360f4446fb9e512c75e7244ab482ff7bf25bc2abc7c65d95302bf1f Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.905142 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.923649 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.936924 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.953379 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.967069 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.967101 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.967110 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.967125 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.967135 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:14Z","lastTransitionTime":"2026-02-18T15:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:14 crc kubenswrapper[4968]: I0218 15:22:14.977951 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:14Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.014521 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.055482 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.069791 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.069840 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.069851 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.069870 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.069890 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:15Z","lastTransitionTime":"2026-02-18T15:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.096662 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.139111 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.173078 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.173117 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.173128 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.173144 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.173154 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:15Z","lastTransitionTime":"2026-02-18T15:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.177324 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 06:56:29.130660746 +0000 UTC Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.179597 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.217363 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.229878 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:15 crc kubenswrapper[4968]: E0218 15:22:15.230019 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.276178 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.276225 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.276236 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.276253 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.276262 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:15Z","lastTransitionTime":"2026-02-18T15:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.379517 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.379556 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.379564 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.379579 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.379588 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:15Z","lastTransitionTime":"2026-02-18T15:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.482521 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.482576 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.482588 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.482606 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.482619 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:15Z","lastTransitionTime":"2026-02-18T15:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.529433 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" event={"ID":"cd365caa-2db2-407a-b264-d18a0d06fe06","Type":"ContainerStarted","Data":"b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.529535 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" event={"ID":"cd365caa-2db2-407a-b264-d18a0d06fe06","Type":"ContainerStarted","Data":"0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.529563 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" event={"ID":"cd365caa-2db2-407a-b264-d18a0d06fe06","Type":"ContainerStarted","Data":"410d93560360f4446fb9e512c75e7244ab482ff7bf25bc2abc7c65d95302bf1f"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.532856 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/1.log" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.533601 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/0.log" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.537741 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7" exitCode=1 Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.537811 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.537914 4968 scope.go:117] "RemoveContainer" containerID="ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.538543 4968 scope.go:117] "RemoveContainer" containerID="196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7" Feb 18 15:22:15 crc kubenswrapper[4968]: E0218 15:22:15.538722 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.552803 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.572229 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.584874 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.584922 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.584938 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.584956 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.584966 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:15Z","lastTransitionTime":"2026-02-18T15:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.587123 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.603586 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.614583 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.627140 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.641409 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.656391 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.670217 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.676226 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-7sk6k"] Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.676735 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:15 crc kubenswrapper[4968]: E0218 15:22:15.676847 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.681084 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.687151 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.687183 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.687194 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.687212 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.687226 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:15Z","lastTransitionTime":"2026-02-18T15:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.690368 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.702417 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.728544 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.728628 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dcz8\" (UniqueName: \"kubernetes.io/projected/9185b44b-bac7-458b-b4d9-4c389da07c14-kube-api-access-5dcz8\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.745153 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.774104 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.789482 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.789527 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.789537 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.789555 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.789568 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:15Z","lastTransitionTime":"2026-02-18T15:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.825577 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"message\\\":\\\"ry.go:160\\\\nI0218 15:22:12.837603 6165 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.837842 6165 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 15:22:12.837960 6165 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.838307 6165 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.838537 6165 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.850893 6165 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0218 15:22:12.850942 6165 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:12.850946 6165 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:12.850970 6165 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:12.850977 6165 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 15:22:12.850983 6165 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:12.863924 6165 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.829414 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dcz8\" (UniqueName: \"kubernetes.io/projected/9185b44b-bac7-458b-b4d9-4c389da07c14-kube-api-access-5dcz8\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.829543 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:15 crc kubenswrapper[4968]: E0218 15:22:15.829631 4968 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:15 crc kubenswrapper[4968]: E0218 15:22:15.829699 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs podName:9185b44b-bac7-458b-b4d9-4c389da07c14 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:16.329681574 +0000 UTC m=+35.715126436 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs") pod "network-metrics-daemon-7sk6k" (UID: "9185b44b-bac7-458b-b4d9-4c389da07c14") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.855437 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.883996 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dcz8\" (UniqueName: \"kubernetes.io/projected/9185b44b-bac7-458b-b4d9-4c389da07c14-kube-api-access-5dcz8\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.892286 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.892324 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.892332 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.892352 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.892363 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:15Z","lastTransitionTime":"2026-02-18T15:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.916134 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.954518 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:15Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.995101 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.995164 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.995180 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.995207 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:15 crc kubenswrapper[4968]: I0218 15:22:15.995225 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:15Z","lastTransitionTime":"2026-02-18T15:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.007219 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.037694 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.086563 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"message\\\":\\\"ry.go:160\\\\nI0218 15:22:12.837603 6165 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.837842 6165 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 15:22:12.837960 6165 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.838307 6165 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.838537 6165 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.850893 6165 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0218 15:22:12.850942 6165 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:12.850946 6165 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:12.850970 6165 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:12.850977 6165 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 15:22:12.850983 6165 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:12.863924 6165 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"tring{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8888, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 15:22:14.635453 6386 services_controller.go:452] Built service openshift-ingress-canary/ingress-canary per-node LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635461 6386 services_controller.go:453] Built service openshift-ingress-canary/ingress-canary template LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635468 6386 services_controller.go:454] Service openshift-ingress-canary/ingress-canary for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF0218 15:22:14.634795 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.099263 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.099317 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.099331 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.099353 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.099368 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:16Z","lastTransitionTime":"2026-02-18T15:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.118982 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.158712 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.177666 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 14:23:33.040080641 +0000 UTC Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.196190 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.203060 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.203105 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.203121 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.203147 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.203160 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:16Z","lastTransitionTime":"2026-02-18T15:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.229884 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:16 crc kubenswrapper[4968]: E0218 15:22:16.230034 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.230126 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:16 crc kubenswrapper[4968]: E0218 15:22:16.230196 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.239817 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.279907 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.306399 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.306474 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.306497 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.306527 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.306548 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:16Z","lastTransitionTime":"2026-02-18T15:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.317351 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.333735 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:16 crc kubenswrapper[4968]: E0218 15:22:16.334019 4968 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:16 crc kubenswrapper[4968]: E0218 15:22:16.334125 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs podName:9185b44b-bac7-458b-b4d9-4c389da07c14 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:17.334090813 +0000 UTC m=+36.719535845 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs") pod "network-metrics-daemon-7sk6k" (UID: "9185b44b-bac7-458b-b4d9-4c389da07c14") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.355975 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.398738 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.409850 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.409885 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.409895 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.409909 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.409918 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:16Z","lastTransitionTime":"2026-02-18T15:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.437587 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.481915 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.512608 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.512664 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.512676 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.512691 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.512702 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:16Z","lastTransitionTime":"2026-02-18T15:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.520912 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.543697 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/1.log" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.556679 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:16Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.614962 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.615026 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.615038 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.615060 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.615079 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:16Z","lastTransitionTime":"2026-02-18T15:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.716996 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.717046 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.717057 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.717076 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.717088 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:16Z","lastTransitionTime":"2026-02-18T15:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.820343 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.820399 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.820416 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.820439 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.820457 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:16Z","lastTransitionTime":"2026-02-18T15:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.923136 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.923199 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.923217 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.923243 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:16 crc kubenswrapper[4968]: I0218 15:22:16.923260 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:16Z","lastTransitionTime":"2026-02-18T15:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.025426 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.025480 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.025492 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.025511 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.025524 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.041293 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.041552 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:22:33.041522666 +0000 UTC m=+52.426967528 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.128149 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.128198 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.128207 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.128224 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.128237 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.142108 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.142141 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.142167 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.142225 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.142291 4968 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.142316 4968 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.142328 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.142370 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.142383 4968 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.142353 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:33.142337442 +0000 UTC m=+52.527782304 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.142426 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:33.142414504 +0000 UTC m=+52.527859366 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.142444 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:33.142435295 +0000 UTC m=+52.527880157 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.142726 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.142894 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.142927 4968 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.143076 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:33.143040512 +0000 UTC m=+52.528485414 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.178681 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 06:02:00.211712759 +0000 UTC Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.229906 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.230084 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.230159 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.230321 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.230818 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.230856 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.230870 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.230892 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.230904 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.332924 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.332994 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.333005 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.333022 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.333040 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.343444 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.343577 4968 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.343623 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs podName:9185b44b-bac7-458b-b4d9-4c389da07c14 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:19.343610303 +0000 UTC m=+38.729055165 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs") pod "network-metrics-daemon-7sk6k" (UID: "9185b44b-bac7-458b-b4d9-4c389da07c14") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.435664 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.435726 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.435741 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.435771 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.435783 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.538592 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.538640 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.538651 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.538669 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.538681 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.641426 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.641466 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.641476 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.641506 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.641517 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.661454 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.661521 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.661531 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.661567 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.661581 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.681373 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:17Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.686735 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.686796 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.686808 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.686824 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.686836 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.705154 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:17Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.709286 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.709325 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.709339 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.709358 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.709372 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.723977 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:17Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.728714 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.728786 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.728801 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.728820 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.728834 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.745002 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:17Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.749472 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.749525 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.749536 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.749553 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.749583 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.766189 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:17Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:17 crc kubenswrapper[4968]: E0218 15:22:17.766332 4968 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.767828 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.767864 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.767878 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.767897 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.767915 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.870731 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.870803 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.870819 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.870836 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.870846 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.974714 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.975168 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.975187 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.975215 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:17 crc kubenswrapper[4968]: I0218 15:22:17.975236 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:17Z","lastTransitionTime":"2026-02-18T15:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.078161 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.078229 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.078246 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.078270 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.078285 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:18Z","lastTransitionTime":"2026-02-18T15:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.179458 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 16:21:03.943203939 +0000 UTC Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.181343 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.181402 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.181428 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.181464 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.181499 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:18Z","lastTransitionTime":"2026-02-18T15:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.230049 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.230079 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:18 crc kubenswrapper[4968]: E0218 15:22:18.230294 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:18 crc kubenswrapper[4968]: E0218 15:22:18.230498 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.284196 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.284235 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.284243 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.284257 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.284269 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:18Z","lastTransitionTime":"2026-02-18T15:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.387514 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.387591 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.387602 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.387619 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.387633 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:18Z","lastTransitionTime":"2026-02-18T15:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.490661 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.490713 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.490723 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.490740 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.490768 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:18Z","lastTransitionTime":"2026-02-18T15:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.593530 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.593575 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.593592 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.593613 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.593631 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:18Z","lastTransitionTime":"2026-02-18T15:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.696309 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.696368 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.696379 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.696398 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.696411 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:18Z","lastTransitionTime":"2026-02-18T15:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.799907 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.799955 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.799967 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.799984 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.799995 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:18Z","lastTransitionTime":"2026-02-18T15:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.903028 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.903095 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.903113 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.903138 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:18 crc kubenswrapper[4968]: I0218 15:22:18.903160 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:18Z","lastTransitionTime":"2026-02-18T15:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.005674 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.005738 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.005772 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.005795 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.005808 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:19Z","lastTransitionTime":"2026-02-18T15:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.108113 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.108149 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.108162 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.108210 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.108223 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:19Z","lastTransitionTime":"2026-02-18T15:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.180417 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 01:10:49.267353567 +0000 UTC Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.210474 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.210515 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.210525 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.210543 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.210557 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:19Z","lastTransitionTime":"2026-02-18T15:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.229983 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.230008 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:19 crc kubenswrapper[4968]: E0218 15:22:19.230153 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:19 crc kubenswrapper[4968]: E0218 15:22:19.230234 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.314204 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.314464 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.314573 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.314697 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.314776 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:19Z","lastTransitionTime":"2026-02-18T15:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.367134 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:19 crc kubenswrapper[4968]: E0218 15:22:19.367367 4968 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:19 crc kubenswrapper[4968]: E0218 15:22:19.367429 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs podName:9185b44b-bac7-458b-b4d9-4c389da07c14 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:23.367410614 +0000 UTC m=+42.752855496 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs") pod "network-metrics-daemon-7sk6k" (UID: "9185b44b-bac7-458b-b4d9-4c389da07c14") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.417504 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.417542 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.417552 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.417567 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.417577 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:19Z","lastTransitionTime":"2026-02-18T15:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.520553 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.520663 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.520689 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.520719 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.520743 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:19Z","lastTransitionTime":"2026-02-18T15:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.623850 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.623892 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.623905 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.623922 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.623934 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:19Z","lastTransitionTime":"2026-02-18T15:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.726453 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.726489 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.726500 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.726514 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.726523 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:19Z","lastTransitionTime":"2026-02-18T15:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.828904 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.828945 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.828956 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.828972 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.828984 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:19Z","lastTransitionTime":"2026-02-18T15:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.931613 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.931683 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.931707 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.931738 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:19 crc kubenswrapper[4968]: I0218 15:22:19.931796 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:19Z","lastTransitionTime":"2026-02-18T15:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.034042 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.034340 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.034404 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.034498 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.034565 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:20Z","lastTransitionTime":"2026-02-18T15:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.137307 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.137375 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.137395 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.137423 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.137441 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:20Z","lastTransitionTime":"2026-02-18T15:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.181223 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 17:37:00.742188826 +0000 UTC Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.230275 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.230311 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:20 crc kubenswrapper[4968]: E0218 15:22:20.230446 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:20 crc kubenswrapper[4968]: E0218 15:22:20.230600 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.239779 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.239825 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.239839 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.239859 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.239874 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:20Z","lastTransitionTime":"2026-02-18T15:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.342159 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.342208 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.342226 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.342248 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.342263 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:20Z","lastTransitionTime":"2026-02-18T15:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.444539 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.444572 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.444591 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.444610 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.444622 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:20Z","lastTransitionTime":"2026-02-18T15:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.547472 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.547524 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.547534 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.547550 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.547561 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:20Z","lastTransitionTime":"2026-02-18T15:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.650508 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.650557 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.650573 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.650596 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.650610 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:20Z","lastTransitionTime":"2026-02-18T15:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.753283 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.753362 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.753375 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.753419 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.753433 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:20Z","lastTransitionTime":"2026-02-18T15:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.856358 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.856402 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.856419 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.856435 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.856443 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:20Z","lastTransitionTime":"2026-02-18T15:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.958893 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.958940 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.958951 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.958969 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:20 crc kubenswrapper[4968]: I0218 15:22:20.958980 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:20Z","lastTransitionTime":"2026-02-18T15:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.061443 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.061503 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.061519 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.061545 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.061561 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:21Z","lastTransitionTime":"2026-02-18T15:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.164376 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.164423 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.164433 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.164451 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.164462 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:21Z","lastTransitionTime":"2026-02-18T15:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.181863 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 14:44:42.416401653 +0000 UTC Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.230391 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.230423 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:21 crc kubenswrapper[4968]: E0218 15:22:21.230549 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:21 crc kubenswrapper[4968]: E0218 15:22:21.230731 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.265110 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.266899 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.266937 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.266949 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.266967 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.266980 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:21Z","lastTransitionTime":"2026-02-18T15:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.283671 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.307009 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec728fec8e16a122c3fd39d1595265156f43d10ac5c26000ba74eecb38524491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"message\\\":\\\"ry.go:160\\\\nI0218 15:22:12.837603 6165 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.837842 6165 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 15:22:12.837960 6165 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.838307 6165 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.838537 6165 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 15:22:12.850893 6165 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0218 15:22:12.850942 6165 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:12.850946 6165 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:12.850970 6165 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:12.850977 6165 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 15:22:12.850983 6165 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:12.863924 6165 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"tring{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8888, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 15:22:14.635453 6386 services_controller.go:452] Built service openshift-ingress-canary/ingress-canary per-node LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635461 6386 services_controller.go:453] Built service openshift-ingress-canary/ingress-canary template LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635468 6386 services_controller.go:454] Service openshift-ingress-canary/ingress-canary for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF0218 15:22:14.634795 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.320538 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.334255 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.348950 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.360909 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.369650 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.369679 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.369688 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.369703 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.369714 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:21Z","lastTransitionTime":"2026-02-18T15:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.377968 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.396647 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.412571 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.425703 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.439369 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.450918 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.460254 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.469660 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.471273 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.471319 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.471333 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.471354 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.471366 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:21Z","lastTransitionTime":"2026-02-18T15:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.483304 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.494065 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:21Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.573333 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.573370 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.573379 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.573395 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.573405 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:21Z","lastTransitionTime":"2026-02-18T15:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.676148 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.676200 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.676212 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.676230 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.676251 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:21Z","lastTransitionTime":"2026-02-18T15:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.778427 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.778462 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.778472 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.778485 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.778495 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:21Z","lastTransitionTime":"2026-02-18T15:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.881057 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.881099 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.881111 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.881127 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.881140 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:21Z","lastTransitionTime":"2026-02-18T15:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.984166 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.984203 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.984213 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.984228 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:21 crc kubenswrapper[4968]: I0218 15:22:21.984238 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:21Z","lastTransitionTime":"2026-02-18T15:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.086675 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.086762 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.086775 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.086792 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.086806 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:22Z","lastTransitionTime":"2026-02-18T15:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.182597 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 01:15:17.389405466 +0000 UTC Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.189764 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.189817 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.189835 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.189854 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.189865 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:22Z","lastTransitionTime":"2026-02-18T15:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.230424 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.230737 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:22 crc kubenswrapper[4968]: E0218 15:22:22.230648 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:22 crc kubenswrapper[4968]: E0218 15:22:22.230978 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.292287 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.292323 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.292331 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.292347 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.292357 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:22Z","lastTransitionTime":"2026-02-18T15:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.395105 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.395158 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.395167 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.395183 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.395195 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:22Z","lastTransitionTime":"2026-02-18T15:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.498216 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.498316 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.498334 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.498390 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.498407 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:22Z","lastTransitionTime":"2026-02-18T15:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.600343 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.600375 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.600385 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.600401 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.600412 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:22Z","lastTransitionTime":"2026-02-18T15:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.702642 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.702685 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.702694 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.702710 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.702719 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:22Z","lastTransitionTime":"2026-02-18T15:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.806287 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.806353 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.806374 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.806397 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.806416 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:22Z","lastTransitionTime":"2026-02-18T15:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.908862 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.908911 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.908928 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.908945 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:22 crc kubenswrapper[4968]: I0218 15:22:22.908958 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:22Z","lastTransitionTime":"2026-02-18T15:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.011838 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.011872 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.011882 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.011897 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.011908 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:23Z","lastTransitionTime":"2026-02-18T15:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.114320 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.114366 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.114379 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.114398 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.114413 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:23Z","lastTransitionTime":"2026-02-18T15:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.183570 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 00:18:28.985941664 +0000 UTC Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.217706 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.217863 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.217896 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.217915 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.217935 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:23Z","lastTransitionTime":"2026-02-18T15:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.230588 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.230659 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:23 crc kubenswrapper[4968]: E0218 15:22:23.230885 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:23 crc kubenswrapper[4968]: E0218 15:22:23.231067 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.321108 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.321626 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.321823 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.321933 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.322020 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:23Z","lastTransitionTime":"2026-02-18T15:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.425092 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.425139 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.425155 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.425175 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.425189 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:23Z","lastTransitionTime":"2026-02-18T15:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.437696 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:23 crc kubenswrapper[4968]: E0218 15:22:23.437897 4968 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:23 crc kubenswrapper[4968]: E0218 15:22:23.437965 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs podName:9185b44b-bac7-458b-b4d9-4c389da07c14 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:31.437947499 +0000 UTC m=+50.823392371 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs") pod "network-metrics-daemon-7sk6k" (UID: "9185b44b-bac7-458b-b4d9-4c389da07c14") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.527844 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.527907 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.527925 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.527945 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.527957 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:23Z","lastTransitionTime":"2026-02-18T15:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.630255 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.630299 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.630309 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.630326 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.630337 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:23Z","lastTransitionTime":"2026-02-18T15:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.732948 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.732990 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.733011 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.733031 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.733047 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:23Z","lastTransitionTime":"2026-02-18T15:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.835329 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.835388 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.835396 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.835411 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.835435 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:23Z","lastTransitionTime":"2026-02-18T15:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.937891 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.937924 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.937932 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.937945 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:23 crc kubenswrapper[4968]: I0218 15:22:23.937955 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:23Z","lastTransitionTime":"2026-02-18T15:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.040397 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.040451 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.040468 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.040491 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.040508 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:24Z","lastTransitionTime":"2026-02-18T15:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.143630 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.143689 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.143704 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.143727 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.143768 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:24Z","lastTransitionTime":"2026-02-18T15:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.184421 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 16:12:04.647109138 +0000 UTC Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.229803 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.229867 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:24 crc kubenswrapper[4968]: E0218 15:22:24.229956 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:24 crc kubenswrapper[4968]: E0218 15:22:24.230062 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.247464 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.247532 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.247550 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.247568 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.247588 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:24Z","lastTransitionTime":"2026-02-18T15:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.350601 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.350660 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.350673 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.350691 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.350703 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:24Z","lastTransitionTime":"2026-02-18T15:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.453331 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.453383 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.453391 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.453410 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.453419 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:24Z","lastTransitionTime":"2026-02-18T15:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.554523 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.555886 4968 scope.go:117] "RemoveContainer" containerID="196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7" Feb 18 15:22:24 crc kubenswrapper[4968]: E0218 15:22:24.556159 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.557206 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.557252 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.557261 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.557277 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.557286 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:24Z","lastTransitionTime":"2026-02-18T15:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.577619 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.608317 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.627879 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.660502 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"tring{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8888, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 15:22:14.635453 6386 services_controller.go:452] Built service openshift-ingress-canary/ingress-canary per-node LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635461 6386 services_controller.go:453] Built service openshift-ingress-canary/ingress-canary template LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635468 6386 services_controller.go:454] Service openshift-ingress-canary/ingress-canary for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF0218 15:22:14.634795 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.661429 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.661485 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.661496 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.661539 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.661552 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:24Z","lastTransitionTime":"2026-02-18T15:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.681697 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.692678 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.705293 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.718518 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.735357 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.754205 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.763911 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.763961 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.763980 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.764001 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.764016 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:24Z","lastTransitionTime":"2026-02-18T15:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.771164 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.788366 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.803617 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.815412 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.826948 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.840343 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.856654 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:24Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.866199 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.866235 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.866246 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.866265 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.866278 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:24Z","lastTransitionTime":"2026-02-18T15:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.968490 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.968549 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.968568 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.968587 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:24 crc kubenswrapper[4968]: I0218 15:22:24.968600 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:24Z","lastTransitionTime":"2026-02-18T15:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.071367 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.071415 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.071429 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.071448 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.071460 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:25Z","lastTransitionTime":"2026-02-18T15:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.174640 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.174700 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.174714 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.174733 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.174766 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:25Z","lastTransitionTime":"2026-02-18T15:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.185146 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 09:57:39.26805377 +0000 UTC Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.230147 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.230162 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:25 crc kubenswrapper[4968]: E0218 15:22:25.230359 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:25 crc kubenswrapper[4968]: E0218 15:22:25.230520 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.277538 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.277605 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.277619 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.277635 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.277655 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:25Z","lastTransitionTime":"2026-02-18T15:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.381054 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.381089 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.381098 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.381112 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.381121 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:25Z","lastTransitionTime":"2026-02-18T15:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.484802 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.484872 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.484882 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.484902 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.484913 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:25Z","lastTransitionTime":"2026-02-18T15:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.589913 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.589963 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.589974 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.589996 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.590009 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:25Z","lastTransitionTime":"2026-02-18T15:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.692880 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.692932 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.692942 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.692962 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.692974 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:25Z","lastTransitionTime":"2026-02-18T15:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.795803 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.795897 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.795915 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.795941 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.795959 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:25Z","lastTransitionTime":"2026-02-18T15:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.899592 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.899672 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.899691 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.899723 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:25 crc kubenswrapper[4968]: I0218 15:22:25.899776 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:25Z","lastTransitionTime":"2026-02-18T15:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.003052 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.003129 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.003148 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.003180 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.003200 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:26Z","lastTransitionTime":"2026-02-18T15:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.106650 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.106734 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.106802 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.106831 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.106849 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:26Z","lastTransitionTime":"2026-02-18T15:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.185794 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 03:32:35.240818412 +0000 UTC Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.209841 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.209912 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.209930 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.209956 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.209975 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:26Z","lastTransitionTime":"2026-02-18T15:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.230598 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.230720 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:26 crc kubenswrapper[4968]: E0218 15:22:26.230860 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:26 crc kubenswrapper[4968]: E0218 15:22:26.230978 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.312675 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.312798 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.312878 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.312917 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.312941 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:26Z","lastTransitionTime":"2026-02-18T15:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.416546 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.416610 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.416636 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.416663 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.416681 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:26Z","lastTransitionTime":"2026-02-18T15:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.519693 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.519825 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.519852 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.519876 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.519892 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:26Z","lastTransitionTime":"2026-02-18T15:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.622810 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.622921 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.622951 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.622986 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.623010 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:26Z","lastTransitionTime":"2026-02-18T15:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.725774 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.725828 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.725843 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.725865 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.725880 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:26Z","lastTransitionTime":"2026-02-18T15:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.828740 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.828826 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.828842 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.828861 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.828876 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:26Z","lastTransitionTime":"2026-02-18T15:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.932339 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.932392 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.932407 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.932425 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:26 crc kubenswrapper[4968]: I0218 15:22:26.932436 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:26Z","lastTransitionTime":"2026-02-18T15:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.034217 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.034269 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.034278 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.034293 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.034305 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:27Z","lastTransitionTime":"2026-02-18T15:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.137417 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.137474 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.137489 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.137512 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.137527 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:27Z","lastTransitionTime":"2026-02-18T15:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.186695 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 07:06:37.996166 +0000 UTC Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.230819 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.230979 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:27 crc kubenswrapper[4968]: E0218 15:22:27.231033 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:27 crc kubenswrapper[4968]: E0218 15:22:27.231211 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.240666 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.240714 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.240724 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.240741 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.240784 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:27Z","lastTransitionTime":"2026-02-18T15:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.343272 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.343333 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.343352 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.343377 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.343396 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:27Z","lastTransitionTime":"2026-02-18T15:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.445733 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.445796 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.445807 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.445826 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.445842 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:27Z","lastTransitionTime":"2026-02-18T15:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.549029 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.549100 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.549114 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.549133 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.549147 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:27Z","lastTransitionTime":"2026-02-18T15:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.651722 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.651838 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.651863 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.651895 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.651916 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:27Z","lastTransitionTime":"2026-02-18T15:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.755358 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.755432 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.755463 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.755527 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.755548 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:27Z","lastTransitionTime":"2026-02-18T15:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.858330 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.858386 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.858397 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.858420 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.858432 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:27Z","lastTransitionTime":"2026-02-18T15:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.962975 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.963033 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.963050 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.963077 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:27 crc kubenswrapper[4968]: I0218 15:22:27.963096 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:27Z","lastTransitionTime":"2026-02-18T15:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.066978 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.067109 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.067147 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.067182 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.067207 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.160442 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.160500 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.160513 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.160531 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.160542 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: E0218 15:22:28.177873 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:28Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.184656 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.184722 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.184787 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.184824 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.184848 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.187934 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 17:17:45.179155189 +0000 UTC Feb 18 15:22:28 crc kubenswrapper[4968]: E0218 15:22:28.207317 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:28Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.213042 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.213099 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.213114 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.213137 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.213153 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.229636 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.229705 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:28 crc kubenswrapper[4968]: E0218 15:22:28.229888 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:28 crc kubenswrapper[4968]: E0218 15:22:28.230120 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:28 crc kubenswrapper[4968]: E0218 15:22:28.233607 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:28Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.239442 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.239495 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.239516 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.239543 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.239559 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: E0218 15:22:28.258329 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:28Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.265208 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.265260 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.265274 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.265294 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.265307 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: E0218 15:22:28.284110 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:28Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:28 crc kubenswrapper[4968]: E0218 15:22:28.284260 4968 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.286784 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.286830 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.286851 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.286876 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.286924 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.390374 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.390422 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.390434 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.390450 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.390459 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.493724 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.493816 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.493831 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.493852 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.493866 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.597159 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.597204 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.597214 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.597234 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.597303 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.700572 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.700623 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.700634 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.700665 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.700682 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.802887 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.802937 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.802948 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.802974 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.802987 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.904716 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.904777 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.904809 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.904824 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:28 crc kubenswrapper[4968]: I0218 15:22:28.904834 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:28Z","lastTransitionTime":"2026-02-18T15:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.007401 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.007475 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.007489 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.007513 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.007537 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:29Z","lastTransitionTime":"2026-02-18T15:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.110206 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.110259 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.110268 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.110283 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.110294 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:29Z","lastTransitionTime":"2026-02-18T15:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.188452 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 22:24:18.706581004 +0000 UTC Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.213602 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.213681 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.213717 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.213781 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.213806 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:29Z","lastTransitionTime":"2026-02-18T15:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.230267 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.230406 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:29 crc kubenswrapper[4968]: E0218 15:22:29.230453 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:29 crc kubenswrapper[4968]: E0218 15:22:29.230630 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.317416 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.317452 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.317462 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.317480 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.317491 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:29Z","lastTransitionTime":"2026-02-18T15:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.420508 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.420565 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.420583 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.420604 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.420618 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:29Z","lastTransitionTime":"2026-02-18T15:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.433663 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.448268 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.456404 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"tring{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8888, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 15:22:14.635453 6386 services_controller.go:452] Built service openshift-ingress-canary/ingress-canary per-node LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635461 6386 services_controller.go:453] Built service openshift-ingress-canary/ingress-canary template LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635468 6386 services_controller.go:454] Service openshift-ingress-canary/ingress-canary for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF0218 15:22:14.634795 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.472401 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.484069 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.508055 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.525970 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.526008 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.526021 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.526037 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.526047 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:29Z","lastTransitionTime":"2026-02-18T15:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.526369 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.547258 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.567473 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.584732 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.604990 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.624181 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.629722 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.629802 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.629820 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.629845 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.629863 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:29Z","lastTransitionTime":"2026-02-18T15:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.641490 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.657070 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.674673 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.690065 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.704787 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.720582 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.734081 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.734140 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.734156 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.734178 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.734192 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:29Z","lastTransitionTime":"2026-02-18T15:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.735878 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:29Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.837265 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.837328 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.837344 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.837366 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.837379 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:29Z","lastTransitionTime":"2026-02-18T15:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.940872 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.940935 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.940948 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.940971 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:29 crc kubenswrapper[4968]: I0218 15:22:29.940984 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:29Z","lastTransitionTime":"2026-02-18T15:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.044962 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.045040 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.045060 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.045085 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.045103 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:30Z","lastTransitionTime":"2026-02-18T15:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.148768 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.148812 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.148824 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.148842 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.148855 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:30Z","lastTransitionTime":"2026-02-18T15:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.188845 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 18:19:42.808121441 +0000 UTC Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.230501 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.230513 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:30 crc kubenswrapper[4968]: E0218 15:22:30.230710 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:30 crc kubenswrapper[4968]: E0218 15:22:30.230968 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.251997 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.252036 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.252051 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.252069 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.252081 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:30Z","lastTransitionTime":"2026-02-18T15:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.355652 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.355720 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.355770 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.355804 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.355822 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:30Z","lastTransitionTime":"2026-02-18T15:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.459240 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.459315 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.459330 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.459355 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.459368 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:30Z","lastTransitionTime":"2026-02-18T15:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.562701 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.562802 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.562817 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.562861 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.562874 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:30Z","lastTransitionTime":"2026-02-18T15:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.666727 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.666795 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.666806 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.666828 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.666838 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:30Z","lastTransitionTime":"2026-02-18T15:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.770350 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.770428 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.770455 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.770486 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.770506 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:30Z","lastTransitionTime":"2026-02-18T15:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.874413 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.874466 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.874480 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.874501 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.874517 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:30Z","lastTransitionTime":"2026-02-18T15:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.976791 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.976828 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.976837 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.976853 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:30 crc kubenswrapper[4968]: I0218 15:22:30.976865 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:30Z","lastTransitionTime":"2026-02-18T15:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.079705 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.079804 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.079817 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.079835 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.079847 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:31Z","lastTransitionTime":"2026-02-18T15:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.183438 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.183514 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.183530 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.183554 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.183568 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:31Z","lastTransitionTime":"2026-02-18T15:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.189726 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 15:27:16.871545842 +0000 UTC Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.230244 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.230336 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:31 crc kubenswrapper[4968]: E0218 15:22:31.230425 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:31 crc kubenswrapper[4968]: E0218 15:22:31.230523 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.252851 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.269194 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.286462 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.286573 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.286614 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.286630 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.286655 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.286676 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:31Z","lastTransitionTime":"2026-02-18T15:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.306276 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.319162 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.330670 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.346568 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.364017 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.378300 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.390212 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.390263 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.390273 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.390292 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.390303 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:31Z","lastTransitionTime":"2026-02-18T15:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.394849 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.407648 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.419969 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.431581 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.443265 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.466052 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.479017 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.492960 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.493002 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.493014 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.493034 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.493051 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:31Z","lastTransitionTime":"2026-02-18T15:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.501911 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"tring{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8888, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 15:22:14.635453 6386 services_controller.go:452] Built service openshift-ingress-canary/ingress-canary per-node LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635461 6386 services_controller.go:453] Built service openshift-ingress-canary/ingress-canary template LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635468 6386 services_controller.go:454] Service openshift-ingress-canary/ingress-canary for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF0218 15:22:14.634795 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.517268 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:31Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.536301 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:31 crc kubenswrapper[4968]: E0218 15:22:31.536566 4968 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:31 crc kubenswrapper[4968]: E0218 15:22:31.536670 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs podName:9185b44b-bac7-458b-b4d9-4c389da07c14 nodeName:}" failed. No retries permitted until 2026-02-18 15:22:47.536645258 +0000 UTC m=+66.922090120 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs") pod "network-metrics-daemon-7sk6k" (UID: "9185b44b-bac7-458b-b4d9-4c389da07c14") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.597165 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.597231 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.597243 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.597263 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.597280 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:31Z","lastTransitionTime":"2026-02-18T15:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.700143 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.700222 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.700239 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.700266 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.700283 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:31Z","lastTransitionTime":"2026-02-18T15:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.803145 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.803209 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.803221 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.803240 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.803253 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:31Z","lastTransitionTime":"2026-02-18T15:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.905965 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.906033 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.906044 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.906061 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:31 crc kubenswrapper[4968]: I0218 15:22:31.906071 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:31Z","lastTransitionTime":"2026-02-18T15:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.009573 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.009637 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.009655 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.009677 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.009692 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:32Z","lastTransitionTime":"2026-02-18T15:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.113088 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.113187 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.113207 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.113236 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.113257 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:32Z","lastTransitionTime":"2026-02-18T15:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.190576 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 06:32:29.884641187 +0000 UTC Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.216688 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.216807 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.216836 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.216870 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.216893 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:32Z","lastTransitionTime":"2026-02-18T15:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.230404 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.230437 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:32 crc kubenswrapper[4968]: E0218 15:22:32.230646 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:32 crc kubenswrapper[4968]: E0218 15:22:32.230713 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.319860 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.319910 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.319918 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.319934 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.319945 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:32Z","lastTransitionTime":"2026-02-18T15:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.423546 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.423604 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.423613 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.423634 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.423645 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:32Z","lastTransitionTime":"2026-02-18T15:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.527539 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.527597 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.527610 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.527631 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.527645 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:32Z","lastTransitionTime":"2026-02-18T15:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.632109 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.632170 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.632183 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.632210 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.632224 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:32Z","lastTransitionTime":"2026-02-18T15:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.735486 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.735544 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.735554 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.735572 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.735585 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:32Z","lastTransitionTime":"2026-02-18T15:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.838858 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.838918 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.838937 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.838966 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.838982 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:32Z","lastTransitionTime":"2026-02-18T15:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.942415 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.942461 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.942471 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.942490 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:32 crc kubenswrapper[4968]: I0218 15:22:32.942500 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:32Z","lastTransitionTime":"2026-02-18T15:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.045541 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.045620 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.045635 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.045661 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.045674 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:33Z","lastTransitionTime":"2026-02-18T15:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.059074 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.059269 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:23:05.059240241 +0000 UTC m=+84.444685103 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.148779 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.148834 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.148848 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.148868 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.148883 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:33Z","lastTransitionTime":"2026-02-18T15:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.160969 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.161060 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.161111 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.161185 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.161213 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.161212 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.161226 4968 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.161298 4968 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.161373 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:23:05.161349795 +0000 UTC m=+84.546794657 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.161383 4968 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.161423 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 15:23:05.161384336 +0000 UTC m=+84.546829408 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.161480 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:23:05.161456738 +0000 UTC m=+84.546901800 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.161388 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.161870 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.161911 4968 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.162047 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 15:23:05.162013494 +0000 UTC m=+84.547458366 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.191362 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 17:01:56.73229195 +0000 UTC Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.229990 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.230148 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.230293 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:33 crc kubenswrapper[4968]: E0218 15:22:33.230354 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.250991 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.251082 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.251100 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.251168 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.251188 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:33Z","lastTransitionTime":"2026-02-18T15:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.356029 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.356104 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.356121 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.356155 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.356176 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:33Z","lastTransitionTime":"2026-02-18T15:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.459338 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.459396 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.459406 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.459426 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.459440 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:33Z","lastTransitionTime":"2026-02-18T15:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.561780 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.561867 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.561877 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.561895 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.561905 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:33Z","lastTransitionTime":"2026-02-18T15:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.665100 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.665173 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.665195 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.665224 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.665247 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:33Z","lastTransitionTime":"2026-02-18T15:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.768581 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.768629 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.768646 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.768664 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.768677 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:33Z","lastTransitionTime":"2026-02-18T15:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.873230 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.873318 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.873343 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.873372 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.873388 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:33Z","lastTransitionTime":"2026-02-18T15:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.977473 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.977564 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.977592 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.977627 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:33 crc kubenswrapper[4968]: I0218 15:22:33.977653 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:33Z","lastTransitionTime":"2026-02-18T15:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.080772 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.080815 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.080825 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.080844 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.080854 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:34Z","lastTransitionTime":"2026-02-18T15:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.184375 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.184541 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.184580 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.184614 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.184634 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:34Z","lastTransitionTime":"2026-02-18T15:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.191574 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 22:21:20.466841986 +0000 UTC Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.232027 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.232078 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:34 crc kubenswrapper[4968]: E0218 15:22:34.232237 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:34 crc kubenswrapper[4968]: E0218 15:22:34.232500 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.289183 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.289246 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.289263 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.289287 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.289305 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:34Z","lastTransitionTime":"2026-02-18T15:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.394133 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.394461 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.394511 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.394549 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.394575 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:34Z","lastTransitionTime":"2026-02-18T15:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.498525 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.498616 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.498641 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.498677 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.498707 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:34Z","lastTransitionTime":"2026-02-18T15:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.601835 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.601881 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.601901 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.601921 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.601932 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:34Z","lastTransitionTime":"2026-02-18T15:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.705167 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.705241 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.705255 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.705278 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.705291 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:34Z","lastTransitionTime":"2026-02-18T15:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.808298 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.808344 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.808358 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.808381 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.808395 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:34Z","lastTransitionTime":"2026-02-18T15:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.911519 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.911571 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.911583 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.911600 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:34 crc kubenswrapper[4968]: I0218 15:22:34.911611 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:34Z","lastTransitionTime":"2026-02-18T15:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.014477 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.014546 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.014622 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.014646 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.014658 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:35Z","lastTransitionTime":"2026-02-18T15:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.117797 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.117856 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.117867 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.117889 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.117902 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:35Z","lastTransitionTime":"2026-02-18T15:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.192140 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 18:58:56.924417153 +0000 UTC Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.220586 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.220994 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.221157 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.221274 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.221359 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:35Z","lastTransitionTime":"2026-02-18T15:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.230043 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.230094 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:35 crc kubenswrapper[4968]: E0218 15:22:35.230604 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:35 crc kubenswrapper[4968]: E0218 15:22:35.230721 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.324947 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.325002 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.325022 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.325046 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.325057 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:35Z","lastTransitionTime":"2026-02-18T15:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.428635 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.428694 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.428709 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.428731 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.428765 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:35Z","lastTransitionTime":"2026-02-18T15:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.532284 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.532349 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.532369 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.532400 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.532420 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:35Z","lastTransitionTime":"2026-02-18T15:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.635949 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.636036 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.636072 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.636114 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.636141 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:35Z","lastTransitionTime":"2026-02-18T15:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.739328 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.739389 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.739409 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.739435 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.739453 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:35Z","lastTransitionTime":"2026-02-18T15:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.843113 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.843199 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.843224 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.843262 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.843286 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:35Z","lastTransitionTime":"2026-02-18T15:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.947900 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.947999 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.948023 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.948062 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:35 crc kubenswrapper[4968]: I0218 15:22:35.948087 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:35Z","lastTransitionTime":"2026-02-18T15:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.051723 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.051838 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.051858 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.051889 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.051909 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:36Z","lastTransitionTime":"2026-02-18T15:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.156326 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.156423 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.156449 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.156487 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.156515 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:36Z","lastTransitionTime":"2026-02-18T15:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.192310 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 03:19:35.1199562 +0000 UTC Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.229860 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.229937 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:36 crc kubenswrapper[4968]: E0218 15:22:36.230064 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:36 crc kubenswrapper[4968]: E0218 15:22:36.230338 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.260497 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.260569 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.260587 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.260615 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.260635 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:36Z","lastTransitionTime":"2026-02-18T15:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.365192 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.365271 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.365292 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.365321 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.365341 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:36Z","lastTransitionTime":"2026-02-18T15:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.468945 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.469496 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.469518 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.469566 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.469587 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:36Z","lastTransitionTime":"2026-02-18T15:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.573776 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.573849 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.573860 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.573879 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.573892 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:36Z","lastTransitionTime":"2026-02-18T15:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.677400 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.677467 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.677489 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.677523 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.677547 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:36Z","lastTransitionTime":"2026-02-18T15:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.781404 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.781492 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.781515 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.781550 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.781572 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:36Z","lastTransitionTime":"2026-02-18T15:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.884881 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.884964 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.884990 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.885023 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.885050 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:36Z","lastTransitionTime":"2026-02-18T15:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.989191 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.989284 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.989317 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.989369 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:36 crc kubenswrapper[4968]: I0218 15:22:36.989393 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:36Z","lastTransitionTime":"2026-02-18T15:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.093380 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.093468 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.093492 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.093530 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.093555 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:37Z","lastTransitionTime":"2026-02-18T15:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.193490 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 00:12:38.353284412 +0000 UTC Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.197597 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.197663 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.197680 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.197708 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.197726 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:37Z","lastTransitionTime":"2026-02-18T15:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.230831 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.230983 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:37 crc kubenswrapper[4968]: E0218 15:22:37.231141 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:37 crc kubenswrapper[4968]: E0218 15:22:37.231416 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.300640 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.300703 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.300717 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.300739 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.300783 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:37Z","lastTransitionTime":"2026-02-18T15:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.403948 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.404021 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.404036 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.404055 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.404454 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:37Z","lastTransitionTime":"2026-02-18T15:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.508424 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.508487 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.508505 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.508533 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.508555 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:37Z","lastTransitionTime":"2026-02-18T15:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.612048 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.612953 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.612976 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.613000 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.613019 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:37Z","lastTransitionTime":"2026-02-18T15:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.716376 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.716422 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.716432 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.716449 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.716460 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:37Z","lastTransitionTime":"2026-02-18T15:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.818811 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.818865 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.818880 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.818900 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.818915 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:37Z","lastTransitionTime":"2026-02-18T15:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.923403 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.923477 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.923495 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.923524 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:37 crc kubenswrapper[4968]: I0218 15:22:37.923542 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:37Z","lastTransitionTime":"2026-02-18T15:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.026737 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.026799 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.026811 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.026828 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.026838 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.130462 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.130525 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.130544 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.130569 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.130582 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.194833 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 06:01:36.183220979 +0000 UTC Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.229900 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.230004 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:38 crc kubenswrapper[4968]: E0218 15:22:38.230521 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:38 crc kubenswrapper[4968]: E0218 15:22:38.230814 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.230979 4968 scope.go:117] "RemoveContainer" containerID="196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.233404 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.233452 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.233467 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.233487 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.233504 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.337678 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.338166 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.338178 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.338202 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.338215 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.441155 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.441205 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.441219 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.441239 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.441254 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.497218 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.497280 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.497290 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.497317 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.497330 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: E0218 15:22:38.515353 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.518963 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.519010 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.519020 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.519041 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.519051 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: E0218 15:22:38.531169 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.541594 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.541663 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.541678 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.541703 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.541720 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: E0218 15:22:38.554547 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.559036 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.559105 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.559127 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.559157 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.559179 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: E0218 15:22:38.572068 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.576413 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.576465 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.576478 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.576498 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.576510 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: E0218 15:22:38.591793 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: E0218 15:22:38.591953 4968 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.593891 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.593958 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.593976 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.594005 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.594027 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.633810 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/1.log" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.637293 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a"} Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.637953 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.661375 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.680025 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.695591 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.696859 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.696895 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.696907 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.696928 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.696940 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.711499 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.729040 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.741291 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.755949 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.768816 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.780547 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.792213 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.799389 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.799432 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.799443 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.799461 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.799473 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.802915 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.821803 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.834367 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.852799 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"tring{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8888, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 15:22:14.635453 6386 services_controller.go:452] Built service openshift-ingress-canary/ingress-canary per-node LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635461 6386 services_controller.go:453] Built service openshift-ingress-canary/ingress-canary template LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635468 6386 services_controller.go:454] Service openshift-ingress-canary/ingress-canary for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF0218 15:22:14.634795 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.866177 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.876083 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.887627 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.901118 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:38Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.902863 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.902914 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.902927 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.902949 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:38 crc kubenswrapper[4968]: I0218 15:22:38.902965 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:38Z","lastTransitionTime":"2026-02-18T15:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.005876 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.005932 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.005950 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.005971 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.005985 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:39Z","lastTransitionTime":"2026-02-18T15:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.109079 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.109133 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.109144 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.109164 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.109179 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:39Z","lastTransitionTime":"2026-02-18T15:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.195726 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 23:32:53.1946513 +0000 UTC Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.212656 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.212715 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.212731 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.212787 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.212805 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:39Z","lastTransitionTime":"2026-02-18T15:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.229827 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.229838 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:39 crc kubenswrapper[4968]: E0218 15:22:39.230076 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:39 crc kubenswrapper[4968]: E0218 15:22:39.230146 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.316311 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.316345 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.316354 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.316371 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.316381 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:39Z","lastTransitionTime":"2026-02-18T15:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.419875 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.419934 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.419944 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.419963 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.419978 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:39Z","lastTransitionTime":"2026-02-18T15:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.523085 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.523131 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.523142 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.523164 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.523178 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:39Z","lastTransitionTime":"2026-02-18T15:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.626863 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.626963 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.627001 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.627045 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.627070 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:39Z","lastTransitionTime":"2026-02-18T15:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.642724 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/2.log" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.643355 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/1.log" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.646023 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a" exitCode=1 Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.646079 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a"} Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.646130 4968 scope.go:117] "RemoveContainer" containerID="196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.647549 4968 scope.go:117] "RemoveContainer" containerID="ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a" Feb 18 15:22:39 crc kubenswrapper[4968]: E0218 15:22:39.647960 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.661573 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.673081 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.685654 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.697254 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.709380 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.722024 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.729968 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.730031 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.730043 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.730066 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.730080 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:39Z","lastTransitionTime":"2026-02-18T15:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.733523 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.744163 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.756209 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.775586 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ddbec066dabd45f68c10f3c1022212509d75f35c9cd388fe1bc8a8a5c8bf7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"tring{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-canary/ingress-canary\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.34\\\\\\\", Port:8888, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0218 15:22:14.635453 6386 services_controller.go:452] Built service openshift-ingress-canary/ingress-canary per-node LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635461 6386 services_controller.go:453] Built service openshift-ingress-canary/ingress-canary template LB for network=default: []services.LB{}\\\\nI0218 15:22:14.635468 6386 services_controller.go:454] Service openshift-ingress-canary/ingress-canary for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF0218 15:22:14.634795 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:39Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0218 15:22:39.090619 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0218 15:22:39.090655 6647 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI0218 15:22:39.090680 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI0218 15:22:39.090693 6647 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:39.090705 6647 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:39.090602 6647 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0218 15:22:39.090903 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0218 15:22:39.090948 6647 factory.go:656] Stopping watch factory\\\\nI0218 15:22:39.090972 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0218 15:22:39.091376 6647 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0218 15:22:39.091558 6647 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0218 15:22:39.091665 6647 ovnkube.go:599] Stopped ovnkube\\\\nI0218 15:22:39.091737 6647 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 15:22:39.091917 6647 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.787180 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.797077 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.818709 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.830570 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.832353 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.832400 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.832412 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.832431 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.832444 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:39Z","lastTransitionTime":"2026-02-18T15:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.846198 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.860943 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.874207 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.888901 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:39Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.935551 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.935612 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.935631 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.935654 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:39 crc kubenswrapper[4968]: I0218 15:22:39.935671 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:39Z","lastTransitionTime":"2026-02-18T15:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.038676 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.038766 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.038783 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.038808 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.038823 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:40Z","lastTransitionTime":"2026-02-18T15:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.141507 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.141559 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.141574 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.141593 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.141604 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:40Z","lastTransitionTime":"2026-02-18T15:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.196335 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 21:51:44.310164841 +0000 UTC Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.229907 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.229945 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:40 crc kubenswrapper[4968]: E0218 15:22:40.230181 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:40 crc kubenswrapper[4968]: E0218 15:22:40.230263 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.244696 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.244806 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.244824 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.244853 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.244870 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:40Z","lastTransitionTime":"2026-02-18T15:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.348456 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.348543 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.348564 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.348595 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.348618 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:40Z","lastTransitionTime":"2026-02-18T15:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.451431 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.451484 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.451500 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.451522 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.451534 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:40Z","lastTransitionTime":"2026-02-18T15:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.554167 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.554208 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.554224 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.554245 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.554258 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:40Z","lastTransitionTime":"2026-02-18T15:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.654200 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/2.log" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.656785 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.656831 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.656849 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.656876 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.656893 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:40Z","lastTransitionTime":"2026-02-18T15:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.659333 4968 scope.go:117] "RemoveContainer" containerID="ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a" Feb 18 15:22:40 crc kubenswrapper[4968]: E0218 15:22:40.659612 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.678107 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.695367 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.709988 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.723009 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.736805 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.755539 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.760483 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.760516 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.760529 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.760550 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.760565 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:40Z","lastTransitionTime":"2026-02-18T15:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.768854 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.780010 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.790434 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.801186 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.831294 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:39Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0218 15:22:39.090619 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0218 15:22:39.090655 6647 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI0218 15:22:39.090680 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI0218 15:22:39.090693 6647 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:39.090705 6647 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:39.090602 6647 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0218 15:22:39.090903 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0218 15:22:39.090948 6647 factory.go:656] Stopping watch factory\\\\nI0218 15:22:39.090972 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0218 15:22:39.091376 6647 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0218 15:22:39.091558 6647 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0218 15:22:39.091665 6647 ovnkube.go:599] Stopped ovnkube\\\\nI0218 15:22:39.091737 6647 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 15:22:39.091917 6647 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.846270 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.858473 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.863657 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.863703 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.863712 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.863732 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.863776 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:40Z","lastTransitionTime":"2026-02-18T15:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.891398 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.905065 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.919404 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.935445 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.951657 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:40Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.967342 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.967412 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.967427 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.967454 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:40 crc kubenswrapper[4968]: I0218 15:22:40.967468 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:40Z","lastTransitionTime":"2026-02-18T15:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.071483 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.071557 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.071575 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.071598 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.071613 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:41Z","lastTransitionTime":"2026-02-18T15:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.174601 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.174660 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.174676 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.174702 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.174717 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:41Z","lastTransitionTime":"2026-02-18T15:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.197544 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 00:12:25.887442546 +0000 UTC Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.230112 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.230203 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:41 crc kubenswrapper[4968]: E0218 15:22:41.230354 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:41 crc kubenswrapper[4968]: E0218 15:22:41.232423 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.247868 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.265578 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.279152 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.279206 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.279224 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.279249 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.279264 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:41Z","lastTransitionTime":"2026-02-18T15:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.292705 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.321259 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.335363 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.346728 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.358130 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.371542 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.382654 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.382955 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.383023 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.383089 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.383199 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:41Z","lastTransitionTime":"2026-02-18T15:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.387785 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.417612 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.435159 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.459489 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:39Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0218 15:22:39.090619 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0218 15:22:39.090655 6647 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI0218 15:22:39.090680 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI0218 15:22:39.090693 6647 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:39.090705 6647 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:39.090602 6647 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0218 15:22:39.090903 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0218 15:22:39.090948 6647 factory.go:656] Stopping watch factory\\\\nI0218 15:22:39.090972 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0218 15:22:39.091376 6647 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0218 15:22:39.091558 6647 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0218 15:22:39.091665 6647 ovnkube.go:599] Stopped ovnkube\\\\nI0218 15:22:39.091737 6647 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 15:22:39.091917 6647 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.476823 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.487022 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.487058 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.487069 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.487084 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.487094 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:41Z","lastTransitionTime":"2026-02-18T15:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.493682 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.512069 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.526556 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.541520 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.558439 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:41Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.590358 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.590428 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.590444 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.590467 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.590482 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:41Z","lastTransitionTime":"2026-02-18T15:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.694079 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.694132 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.694145 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.694164 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.694178 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:41Z","lastTransitionTime":"2026-02-18T15:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.797016 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.797083 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.797101 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.797125 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.797143 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:41Z","lastTransitionTime":"2026-02-18T15:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.901039 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.901131 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.901166 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.901192 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:41 crc kubenswrapper[4968]: I0218 15:22:41.901206 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:41Z","lastTransitionTime":"2026-02-18T15:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.004852 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.004931 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.004952 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.004982 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.005025 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:42Z","lastTransitionTime":"2026-02-18T15:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.109042 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.109553 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.109566 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.109588 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.109602 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:42Z","lastTransitionTime":"2026-02-18T15:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.198553 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 11:41:05.722051532 +0000 UTC Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.213245 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.213320 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.213334 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.213356 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.213370 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:42Z","lastTransitionTime":"2026-02-18T15:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.231118 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.231136 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:42 crc kubenswrapper[4968]: E0218 15:22:42.231403 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:42 crc kubenswrapper[4968]: E0218 15:22:42.231470 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.317041 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.317092 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.317106 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.317128 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.317142 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:42Z","lastTransitionTime":"2026-02-18T15:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.420523 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.420612 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.420628 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.420706 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.420788 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:42Z","lastTransitionTime":"2026-02-18T15:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.523812 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.523873 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.523883 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.523915 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.523927 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:42Z","lastTransitionTime":"2026-02-18T15:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.626840 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.626890 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.626906 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.626927 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.626941 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:42Z","lastTransitionTime":"2026-02-18T15:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.730559 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.730641 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.730657 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.730678 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.730693 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:42Z","lastTransitionTime":"2026-02-18T15:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.835619 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.835697 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.835715 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.835744 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.835793 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:42Z","lastTransitionTime":"2026-02-18T15:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.939250 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.939326 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.939340 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.939363 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:42 crc kubenswrapper[4968]: I0218 15:22:42.939379 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:42Z","lastTransitionTime":"2026-02-18T15:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.043118 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.043190 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.043207 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.043231 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.043245 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:43Z","lastTransitionTime":"2026-02-18T15:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.146822 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.146881 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.146895 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.146918 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.146933 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:43Z","lastTransitionTime":"2026-02-18T15:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.199840 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 12:24:29.972257519 +0000 UTC Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.230346 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.230460 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:43 crc kubenswrapper[4968]: E0218 15:22:43.230620 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:43 crc kubenswrapper[4968]: E0218 15:22:43.230793 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.250563 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.250615 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.250628 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.250650 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.250663 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:43Z","lastTransitionTime":"2026-02-18T15:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.354154 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.354218 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.354233 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.354254 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.354270 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:43Z","lastTransitionTime":"2026-02-18T15:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.457665 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.457735 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.457768 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.457795 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.457811 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:43Z","lastTransitionTime":"2026-02-18T15:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.561685 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.561731 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.561771 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.561797 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.561814 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:43Z","lastTransitionTime":"2026-02-18T15:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.664833 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.664901 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.664916 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.664938 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.664952 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:43Z","lastTransitionTime":"2026-02-18T15:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.769395 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.769490 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.769504 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.769562 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.769576 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:43Z","lastTransitionTime":"2026-02-18T15:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.872742 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.872827 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.872840 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.872862 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.872875 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:43Z","lastTransitionTime":"2026-02-18T15:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.975994 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.976040 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.976053 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.976073 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:43 crc kubenswrapper[4968]: I0218 15:22:43.976086 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:43Z","lastTransitionTime":"2026-02-18T15:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.078639 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.078682 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.078693 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.078713 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.078725 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:44Z","lastTransitionTime":"2026-02-18T15:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.181623 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.181688 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.181707 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.181730 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.181774 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:44Z","lastTransitionTime":"2026-02-18T15:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.200320 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 03:23:00.650890325 +0000 UTC Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.229630 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.229686 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:44 crc kubenswrapper[4968]: E0218 15:22:44.229825 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:44 crc kubenswrapper[4968]: E0218 15:22:44.230041 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.285696 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.285743 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.285793 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.285817 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.285834 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:44Z","lastTransitionTime":"2026-02-18T15:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.389794 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.389844 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.389860 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.389884 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.389906 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:44Z","lastTransitionTime":"2026-02-18T15:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.493852 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.493922 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.493941 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.493966 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.493984 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:44Z","lastTransitionTime":"2026-02-18T15:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.597099 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.597177 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.597198 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.597229 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.597246 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:44Z","lastTransitionTime":"2026-02-18T15:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.700095 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.700150 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.700162 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.700183 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.700195 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:44Z","lastTransitionTime":"2026-02-18T15:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.804233 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.804302 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.804326 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.804374 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.804393 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:44Z","lastTransitionTime":"2026-02-18T15:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.908133 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.908180 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.908194 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.908214 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:44 crc kubenswrapper[4968]: I0218 15:22:44.908230 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:44Z","lastTransitionTime":"2026-02-18T15:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.011835 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.011877 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.011890 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.011912 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.011927 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:45Z","lastTransitionTime":"2026-02-18T15:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.114428 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.114476 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.114489 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.114512 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.114526 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:45Z","lastTransitionTime":"2026-02-18T15:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.200949 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 15:15:56.564323195 +0000 UTC Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.218376 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.218445 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.218461 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.218485 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.218502 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:45Z","lastTransitionTime":"2026-02-18T15:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.229999 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.230156 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:45 crc kubenswrapper[4968]: E0218 15:22:45.230313 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:45 crc kubenswrapper[4968]: E0218 15:22:45.230478 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.321148 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.321191 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.321202 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.321218 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.321229 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:45Z","lastTransitionTime":"2026-02-18T15:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.424575 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.424617 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.424625 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.424641 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.424651 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:45Z","lastTransitionTime":"2026-02-18T15:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.527328 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.527368 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.527377 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.527393 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.527403 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:45Z","lastTransitionTime":"2026-02-18T15:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.630241 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.630288 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.630296 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.630313 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.630325 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:45Z","lastTransitionTime":"2026-02-18T15:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.733500 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.733576 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.733600 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.733639 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.733668 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:45Z","lastTransitionTime":"2026-02-18T15:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.836985 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.837049 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.837060 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.837081 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.837096 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:45Z","lastTransitionTime":"2026-02-18T15:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.940387 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.940440 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.940451 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.940472 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:45 crc kubenswrapper[4968]: I0218 15:22:45.940488 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:45Z","lastTransitionTime":"2026-02-18T15:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.042858 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.042925 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.042939 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.042961 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.042976 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:46Z","lastTransitionTime":"2026-02-18T15:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.146306 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.146385 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.146403 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.146430 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.146448 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:46Z","lastTransitionTime":"2026-02-18T15:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.202028 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 19:04:52.790131796 +0000 UTC Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.229931 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.230041 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:46 crc kubenswrapper[4968]: E0218 15:22:46.230105 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:46 crc kubenswrapper[4968]: E0218 15:22:46.230290 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.250151 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.250215 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.250230 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.250253 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.250268 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:46Z","lastTransitionTime":"2026-02-18T15:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.353290 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.353336 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.353347 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.353366 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.353381 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:46Z","lastTransitionTime":"2026-02-18T15:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.456888 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.456926 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.456937 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.456953 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.456967 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:46Z","lastTransitionTime":"2026-02-18T15:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.560241 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.560280 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.560289 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.560304 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.560316 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:46Z","lastTransitionTime":"2026-02-18T15:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.662844 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.662889 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.662899 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.662917 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.662929 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:46Z","lastTransitionTime":"2026-02-18T15:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.766786 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.766844 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.766857 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.766878 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.766891 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:46Z","lastTransitionTime":"2026-02-18T15:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.869786 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.869854 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.869867 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.869888 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.869904 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:46Z","lastTransitionTime":"2026-02-18T15:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.973202 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.973243 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.973253 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.973269 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:46 crc kubenswrapper[4968]: I0218 15:22:46.973280 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:46Z","lastTransitionTime":"2026-02-18T15:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.076377 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.076433 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.076446 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.076473 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.076489 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:47Z","lastTransitionTime":"2026-02-18T15:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.179309 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.179358 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.179373 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.179391 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.179403 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:47Z","lastTransitionTime":"2026-02-18T15:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.203265 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 23:38:42.734349876 +0000 UTC Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.230033 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:47 crc kubenswrapper[4968]: E0218 15:22:47.230248 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.230033 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:47 crc kubenswrapper[4968]: E0218 15:22:47.230580 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.282891 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.282934 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.282943 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.282960 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.282970 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:47Z","lastTransitionTime":"2026-02-18T15:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.385691 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.385739 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.385782 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.385804 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.385815 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:47Z","lastTransitionTime":"2026-02-18T15:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.487847 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.487893 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.487906 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.487928 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.487943 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:47Z","lastTransitionTime":"2026-02-18T15:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.590452 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.590500 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.590515 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.590537 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.590583 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:47Z","lastTransitionTime":"2026-02-18T15:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.629902 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:47 crc kubenswrapper[4968]: E0218 15:22:47.630104 4968 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:47 crc kubenswrapper[4968]: E0218 15:22:47.630195 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs podName:9185b44b-bac7-458b-b4d9-4c389da07c14 nodeName:}" failed. No retries permitted until 2026-02-18 15:23:19.630172125 +0000 UTC m=+99.015616987 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs") pod "network-metrics-daemon-7sk6k" (UID: "9185b44b-bac7-458b-b4d9-4c389da07c14") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.693594 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.693638 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.693649 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.693665 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.693678 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:47Z","lastTransitionTime":"2026-02-18T15:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.796397 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.796440 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.796450 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.796467 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.796477 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:47Z","lastTransitionTime":"2026-02-18T15:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.899496 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.899548 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.899561 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.899583 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:47 crc kubenswrapper[4968]: I0218 15:22:47.899597 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:47Z","lastTransitionTime":"2026-02-18T15:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.002264 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.002308 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.002322 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.002342 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.002353 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.105533 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.105570 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.105579 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.105595 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.105606 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.203457 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 08:03:52.186838639 +0000 UTC Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.208776 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.208808 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.208819 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.208836 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.208847 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.229963 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.230041 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:48 crc kubenswrapper[4968]: E0218 15:22:48.230126 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:48 crc kubenswrapper[4968]: E0218 15:22:48.230223 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.311966 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.312081 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.312131 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.312160 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.312175 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.415645 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.415698 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.415710 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.415733 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.415769 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.518876 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.518924 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.518943 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.518962 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.518976 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.622831 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.622926 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.622946 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.622978 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.622996 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.697247 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.697309 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.697319 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.697337 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.697348 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: E0218 15:22:48.710977 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:48Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.715240 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.715302 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.715320 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.715835 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.715865 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: E0218 15:22:48.729537 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:48Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.739645 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.739692 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.739703 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.739720 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.739734 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: E0218 15:22:48.754675 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:48Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.760002 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.760070 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.760119 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.760146 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.760167 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: E0218 15:22:48.773042 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:48Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.776996 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.777040 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.777058 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.777085 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.777098 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: E0218 15:22:48.791641 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:48Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:48 crc kubenswrapper[4968]: E0218 15:22:48.791771 4968 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.793813 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.793873 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.793889 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.793912 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.793925 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.897426 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.897501 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.897515 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.897542 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:48 crc kubenswrapper[4968]: I0218 15:22:48.897557 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:48Z","lastTransitionTime":"2026-02-18T15:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.000777 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.000869 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.000885 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.000904 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.000920 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:49Z","lastTransitionTime":"2026-02-18T15:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.103523 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.103574 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.103585 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.103605 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.103621 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:49Z","lastTransitionTime":"2026-02-18T15:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.204715 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 23:35:28.541737271 +0000 UTC Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.206989 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.207054 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.207072 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.207099 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.207117 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:49Z","lastTransitionTime":"2026-02-18T15:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.230461 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.230502 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:49 crc kubenswrapper[4968]: E0218 15:22:49.230718 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:49 crc kubenswrapper[4968]: E0218 15:22:49.230912 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.310983 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.311057 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.311078 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.311104 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.311122 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:49Z","lastTransitionTime":"2026-02-18T15:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.413708 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.413776 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.413816 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.413846 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.413861 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:49Z","lastTransitionTime":"2026-02-18T15:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.516515 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.516589 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.516608 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.516635 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.516659 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:49Z","lastTransitionTime":"2026-02-18T15:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.620172 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.620242 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.620264 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.620291 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.620309 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:49Z","lastTransitionTime":"2026-02-18T15:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.722586 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.722643 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.722653 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.722670 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.722684 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:49Z","lastTransitionTime":"2026-02-18T15:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.825829 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.825902 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.825918 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.825939 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.825953 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:49Z","lastTransitionTime":"2026-02-18T15:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.929122 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.929174 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.929184 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.929201 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:49 crc kubenswrapper[4968]: I0218 15:22:49.929212 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:49Z","lastTransitionTime":"2026-02-18T15:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.031801 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.031856 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.031866 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.031885 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.031897 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:50Z","lastTransitionTime":"2026-02-18T15:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.136287 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.136344 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.136357 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.136378 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.136396 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:50Z","lastTransitionTime":"2026-02-18T15:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.205490 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 22:43:05.861169613 +0000 UTC Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.230143 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.230221 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:50 crc kubenswrapper[4968]: E0218 15:22:50.230309 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:50 crc kubenswrapper[4968]: E0218 15:22:50.230426 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.239416 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.239463 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.239476 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.239498 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.239513 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:50Z","lastTransitionTime":"2026-02-18T15:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.342151 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.342196 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.342211 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.342230 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.342242 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:50Z","lastTransitionTime":"2026-02-18T15:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.444676 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.444738 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.444791 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.444810 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.444825 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:50Z","lastTransitionTime":"2026-02-18T15:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.547279 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.547360 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.547373 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.547393 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.547404 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:50Z","lastTransitionTime":"2026-02-18T15:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.653450 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.653530 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.653545 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.653570 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.653587 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:50Z","lastTransitionTime":"2026-02-18T15:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.756999 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.757054 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.757066 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.757096 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.757106 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:50Z","lastTransitionTime":"2026-02-18T15:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.859573 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.859637 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.859654 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.859677 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.859696 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:50Z","lastTransitionTime":"2026-02-18T15:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.962325 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.962372 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.962385 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.962402 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:50 crc kubenswrapper[4968]: I0218 15:22:50.962412 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:50Z","lastTransitionTime":"2026-02-18T15:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.065324 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.065366 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.065376 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.065394 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.065405 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:51Z","lastTransitionTime":"2026-02-18T15:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.168134 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.168185 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.168197 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.168217 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.168229 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:51Z","lastTransitionTime":"2026-02-18T15:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.206581 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 01:25:50.789062768 +0000 UTC Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.230351 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.230463 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:51 crc kubenswrapper[4968]: E0218 15:22:51.230723 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:51 crc kubenswrapper[4968]: E0218 15:22:51.230812 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.244838 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.261061 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.271079 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.271127 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.271139 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.271159 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.271170 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:51Z","lastTransitionTime":"2026-02-18T15:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.275993 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.291172 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.304414 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.318929 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.330974 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.341487 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.351980 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.363000 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.373725 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.373782 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.373796 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.373823 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.373839 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:51Z","lastTransitionTime":"2026-02-18T15:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.375313 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.384929 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.394137 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.403277 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.422807 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:39Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0218 15:22:39.090619 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0218 15:22:39.090655 6647 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI0218 15:22:39.090680 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI0218 15:22:39.090693 6647 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:39.090705 6647 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:39.090602 6647 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0218 15:22:39.090903 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0218 15:22:39.090948 6647 factory.go:656] Stopping watch factory\\\\nI0218 15:22:39.090972 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0218 15:22:39.091376 6647 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0218 15:22:39.091558 6647 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0218 15:22:39.091665 6647 ovnkube.go:599] Stopped ovnkube\\\\nI0218 15:22:39.091737 6647 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 15:22:39.091917 6647 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.435653 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.445208 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.466001 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.476664 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.476725 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.476739 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.476777 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.476793 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:51Z","lastTransitionTime":"2026-02-18T15:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.580955 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.581050 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.581074 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.581142 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.581168 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:51Z","lastTransitionTime":"2026-02-18T15:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.684897 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.684956 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.684970 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.684992 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.685005 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:51Z","lastTransitionTime":"2026-02-18T15:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.706671 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2qq8_2d01781a-6a78-49a2-80c7-9ac02c810e3f/kube-multus/0.log" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.706765 4968 generic.go:334] "Generic (PLEG): container finished" podID="2d01781a-6a78-49a2-80c7-9ac02c810e3f" containerID="7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd" exitCode=1 Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.706814 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2qq8" event={"ID":"2d01781a-6a78-49a2-80c7-9ac02c810e3f","Type":"ContainerDied","Data":"7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd"} Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.707289 4968 scope.go:117] "RemoveContainer" containerID="7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.722189 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.735722 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.748525 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.770945 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.785656 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.788376 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.788430 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.788440 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.788460 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.788473 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:51Z","lastTransitionTime":"2026-02-18T15:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.806691 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:39Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0218 15:22:39.090619 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0218 15:22:39.090655 6647 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI0218 15:22:39.090680 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI0218 15:22:39.090693 6647 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:39.090705 6647 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:39.090602 6647 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0218 15:22:39.090903 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0218 15:22:39.090948 6647 factory.go:656] Stopping watch factory\\\\nI0218 15:22:39.090972 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0218 15:22:39.091376 6647 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0218 15:22:39.091558 6647 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0218 15:22:39.091665 6647 ovnkube.go:599] Stopped ovnkube\\\\nI0218 15:22:39.091737 6647 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 15:22:39.091917 6647 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.827830 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:51Z\\\",\\\"message\\\":\\\"2026-02-18T15:22:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae\\\\n2026-02-18T15:22:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae to /host/opt/cni/bin/\\\\n2026-02-18T15:22:06Z [verbose] multus-daemon started\\\\n2026-02-18T15:22:06Z [verbose] Readiness Indicator file check\\\\n2026-02-18T15:22:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.843021 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.856938 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.870226 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.888388 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.891126 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.891168 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.891185 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.891207 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.891218 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:51Z","lastTransitionTime":"2026-02-18T15:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.901975 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.914763 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.929609 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.944334 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.958567 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.973138 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.986155 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:51Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.994152 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.994202 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.994214 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.994234 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:51 crc kubenswrapper[4968]: I0218 15:22:51.994245 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:51Z","lastTransitionTime":"2026-02-18T15:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.096900 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.096963 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.096980 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.097008 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.097029 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:52Z","lastTransitionTime":"2026-02-18T15:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.199149 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.199194 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.199206 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.199224 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.199237 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:52Z","lastTransitionTime":"2026-02-18T15:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.207435 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 03:28:24.336301726 +0000 UTC Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.229873 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.229936 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:52 crc kubenswrapper[4968]: E0218 15:22:52.230305 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:52 crc kubenswrapper[4968]: E0218 15:22:52.230449 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.230809 4968 scope.go:117] "RemoveContainer" containerID="ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a" Feb 18 15:22:52 crc kubenswrapper[4968]: E0218 15:22:52.231150 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.302016 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.302068 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.302082 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.302103 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.302118 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:52Z","lastTransitionTime":"2026-02-18T15:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.404890 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.404930 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.404943 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.404958 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.404968 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:52Z","lastTransitionTime":"2026-02-18T15:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.508213 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.508259 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.508270 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.508294 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.508319 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:52Z","lastTransitionTime":"2026-02-18T15:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.611095 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.611141 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.611155 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.611174 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.611186 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:52Z","lastTransitionTime":"2026-02-18T15:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.712300 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2qq8_2d01781a-6a78-49a2-80c7-9ac02c810e3f/kube-multus/0.log" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.712368 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2qq8" event={"ID":"2d01781a-6a78-49a2-80c7-9ac02c810e3f","Type":"ContainerStarted","Data":"b6ccc51a3ebdf422ba17225bac5935ff711e8ed81619d01a461692c3fb9083eb"} Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.712835 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.712866 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.712880 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.712896 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.712908 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:52Z","lastTransitionTime":"2026-02-18T15:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.729547 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.751191 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:39Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0218 15:22:39.090619 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0218 15:22:39.090655 6647 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI0218 15:22:39.090680 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI0218 15:22:39.090693 6647 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:39.090705 6647 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:39.090602 6647 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0218 15:22:39.090903 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0218 15:22:39.090948 6647 factory.go:656] Stopping watch factory\\\\nI0218 15:22:39.090972 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0218 15:22:39.091376 6647 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0218 15:22:39.091558 6647 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0218 15:22:39.091665 6647 ovnkube.go:599] Stopped ovnkube\\\\nI0218 15:22:39.091737 6647 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 15:22:39.091917 6647 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.765100 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6ccc51a3ebdf422ba17225bac5935ff711e8ed81619d01a461692c3fb9083eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:51Z\\\",\\\"message\\\":\\\"2026-02-18T15:22:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae\\\\n2026-02-18T15:22:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae to /host/opt/cni/bin/\\\\n2026-02-18T15:22:06Z [verbose] multus-daemon started\\\\n2026-02-18T15:22:06Z [verbose] Readiness Indicator file check\\\\n2026-02-18T15:22:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.774351 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.794583 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.808390 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.816079 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.816147 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.816156 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.816174 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.816186 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:52Z","lastTransitionTime":"2026-02-18T15:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.822175 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.833785 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.843206 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.854259 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.865898 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.875675 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.886734 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.897787 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.908575 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.918679 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.918713 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.918760 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.918781 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.918793 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:52Z","lastTransitionTime":"2026-02-18T15:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.921721 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.936322 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:52 crc kubenswrapper[4968]: I0218 15:22:52.947655 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:52Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.022376 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.022930 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.023085 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.023245 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.023652 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:53Z","lastTransitionTime":"2026-02-18T15:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.126484 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.126528 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.126538 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.126554 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.126565 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:53Z","lastTransitionTime":"2026-02-18T15:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.207795 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 14:45:20.579751896 +0000 UTC Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.229378 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.229430 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.229438 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.229456 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.229466 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:53Z","lastTransitionTime":"2026-02-18T15:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.230081 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.230081 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:53 crc kubenswrapper[4968]: E0218 15:22:53.230218 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:53 crc kubenswrapper[4968]: E0218 15:22:53.230367 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.333066 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.333122 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.333136 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.333158 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.333171 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:53Z","lastTransitionTime":"2026-02-18T15:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.436278 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.436350 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.436362 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.436381 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.436393 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:53Z","lastTransitionTime":"2026-02-18T15:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.539172 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.539246 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.539256 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.539283 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.539296 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:53Z","lastTransitionTime":"2026-02-18T15:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.642373 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.642416 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.642426 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.642443 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.642454 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:53Z","lastTransitionTime":"2026-02-18T15:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.745696 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.745770 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.745781 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.745799 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.745810 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:53Z","lastTransitionTime":"2026-02-18T15:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.849627 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.849714 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.849741 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.849806 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.849830 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:53Z","lastTransitionTime":"2026-02-18T15:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.953033 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.953104 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.953126 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.953152 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:53 crc kubenswrapper[4968]: I0218 15:22:53.953170 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:53Z","lastTransitionTime":"2026-02-18T15:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.055555 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.055628 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.055640 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.055665 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.055677 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:54Z","lastTransitionTime":"2026-02-18T15:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.158939 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.159002 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.159019 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.159039 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.159052 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:54Z","lastTransitionTime":"2026-02-18T15:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.208513 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 05:09:37.307380324 +0000 UTC Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.230082 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.230120 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:54 crc kubenswrapper[4968]: E0218 15:22:54.230336 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:54 crc kubenswrapper[4968]: E0218 15:22:54.230404 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.262846 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.262892 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.262901 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.262917 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.262927 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:54Z","lastTransitionTime":"2026-02-18T15:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.366686 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.366994 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.367044 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.367085 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.367101 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:54Z","lastTransitionTime":"2026-02-18T15:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.470007 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.470080 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.470102 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.470131 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.470151 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:54Z","lastTransitionTime":"2026-02-18T15:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.573115 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.573169 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.573183 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.573204 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.573222 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:54Z","lastTransitionTime":"2026-02-18T15:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.676320 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.676386 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.676405 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.676436 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.676457 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:54Z","lastTransitionTime":"2026-02-18T15:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.779951 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.780327 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.780343 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.780372 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.780386 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:54Z","lastTransitionTime":"2026-02-18T15:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.883555 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.883601 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.883612 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.883652 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.883666 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:54Z","lastTransitionTime":"2026-02-18T15:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.986975 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.987039 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.987050 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.987067 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:54 crc kubenswrapper[4968]: I0218 15:22:54.987078 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:54Z","lastTransitionTime":"2026-02-18T15:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.090491 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.090554 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.090566 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.090587 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.090601 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:55Z","lastTransitionTime":"2026-02-18T15:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.194299 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.194368 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.194387 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.194412 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.194430 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:55Z","lastTransitionTime":"2026-02-18T15:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.209776 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 21:37:57.728233438 +0000 UTC Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.230510 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.230510 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:55 crc kubenswrapper[4968]: E0218 15:22:55.230902 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:55 crc kubenswrapper[4968]: E0218 15:22:55.233864 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.246395 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.298009 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.298086 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.298100 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.298118 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.298132 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:55Z","lastTransitionTime":"2026-02-18T15:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.402399 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.402450 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.402465 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.402485 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.402499 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:55Z","lastTransitionTime":"2026-02-18T15:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.506246 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.506371 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.506588 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.506622 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.506645 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:55Z","lastTransitionTime":"2026-02-18T15:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.609961 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.610030 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.610044 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.610066 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.610103 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:55Z","lastTransitionTime":"2026-02-18T15:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.712946 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.713010 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.713029 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.713055 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.713073 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:55Z","lastTransitionTime":"2026-02-18T15:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.815698 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.815791 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.815805 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.815825 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.815838 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:55Z","lastTransitionTime":"2026-02-18T15:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.918801 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.918850 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.918863 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.918880 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:55 crc kubenswrapper[4968]: I0218 15:22:55.918892 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:55Z","lastTransitionTime":"2026-02-18T15:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.022639 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.022715 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.022730 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.022787 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.022807 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:56Z","lastTransitionTime":"2026-02-18T15:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.126067 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.126153 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.126168 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.126191 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.126205 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:56Z","lastTransitionTime":"2026-02-18T15:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.210443 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 14:26:17.434114637 +0000 UTC Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.228952 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.229016 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.229033 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.229058 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.229071 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:56Z","lastTransitionTime":"2026-02-18T15:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.229968 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.230069 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:56 crc kubenswrapper[4968]: E0218 15:22:56.230133 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:56 crc kubenswrapper[4968]: E0218 15:22:56.230294 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.333136 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.333194 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.333208 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.333229 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.333242 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:56Z","lastTransitionTime":"2026-02-18T15:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.436615 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.436708 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.436725 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.437423 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.437497 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:56Z","lastTransitionTime":"2026-02-18T15:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.540096 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.540134 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.540145 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.540160 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.540171 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:56Z","lastTransitionTime":"2026-02-18T15:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.643797 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.643850 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.643863 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.643886 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.643903 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:56Z","lastTransitionTime":"2026-02-18T15:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.760782 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.760871 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.760890 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.760916 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.760935 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:56Z","lastTransitionTime":"2026-02-18T15:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.865127 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.865180 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.865201 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.865227 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.865245 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:56Z","lastTransitionTime":"2026-02-18T15:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.969063 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.969145 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.969173 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.969203 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:56 crc kubenswrapper[4968]: I0218 15:22:56.969223 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:56Z","lastTransitionTime":"2026-02-18T15:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.072622 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.072672 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.072682 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.072698 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.072710 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:57Z","lastTransitionTime":"2026-02-18T15:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.176623 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.176705 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.176728 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.176807 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.176835 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:57Z","lastTransitionTime":"2026-02-18T15:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.211480 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 14:05:59.134873596 +0000 UTC Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.230049 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.230236 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:57 crc kubenswrapper[4968]: E0218 15:22:57.230272 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:57 crc kubenswrapper[4968]: E0218 15:22:57.230522 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.280370 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.280421 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.280456 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.280474 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.280484 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:57Z","lastTransitionTime":"2026-02-18T15:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.382646 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.382774 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.382837 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.382873 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.382896 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:57Z","lastTransitionTime":"2026-02-18T15:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.486209 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.486282 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.486305 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.486335 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.486354 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:57Z","lastTransitionTime":"2026-02-18T15:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.589651 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.589701 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.589715 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.589737 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.589775 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:57Z","lastTransitionTime":"2026-02-18T15:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.693183 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.693243 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.693256 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.693275 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.693289 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:57Z","lastTransitionTime":"2026-02-18T15:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.796641 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.796678 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.796687 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.796707 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.796724 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:57Z","lastTransitionTime":"2026-02-18T15:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.899970 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.900040 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.900065 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.900097 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:57 crc kubenswrapper[4968]: I0218 15:22:57.900120 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:57Z","lastTransitionTime":"2026-02-18T15:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.003153 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.003231 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.003252 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.003283 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.003302 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:58Z","lastTransitionTime":"2026-02-18T15:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.107271 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.107349 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.107368 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.107398 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.107422 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:58Z","lastTransitionTime":"2026-02-18T15:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.211548 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.211638 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.211664 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.211697 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.211689 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 03:38:14.6796915 +0000 UTC Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.211720 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:58Z","lastTransitionTime":"2026-02-18T15:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.230049 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.230112 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:22:58 crc kubenswrapper[4968]: E0218 15:22:58.230242 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:22:58 crc kubenswrapper[4968]: E0218 15:22:58.230409 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.315411 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.315457 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.315467 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.315484 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.315495 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:58Z","lastTransitionTime":"2026-02-18T15:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.418260 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.418328 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.418338 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.418361 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.418390 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:58Z","lastTransitionTime":"2026-02-18T15:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.522357 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.522399 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.522409 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.522473 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.522488 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:58Z","lastTransitionTime":"2026-02-18T15:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.626240 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.626301 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.626311 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.626333 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.626345 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:58Z","lastTransitionTime":"2026-02-18T15:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.729095 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.729177 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.729197 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.729224 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.729243 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:58Z","lastTransitionTime":"2026-02-18T15:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.833347 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.833411 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.833425 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.833451 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.833474 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:58Z","lastTransitionTime":"2026-02-18T15:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.937241 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.937328 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.937362 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.937402 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.937433 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:58Z","lastTransitionTime":"2026-02-18T15:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.989295 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.989343 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.989354 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.989370 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:58 crc kubenswrapper[4968]: I0218 15:22:58.989381 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:58Z","lastTransitionTime":"2026-02-18T15:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: E0218 15:22:59.005224 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:59Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.009196 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.009252 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.009266 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.009290 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.009304 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: E0218 15:22:59.022907 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:59Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.027545 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.027617 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.027631 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.027652 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.027665 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: E0218 15:22:59.039494 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:59Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.043200 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.043235 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.043249 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.043268 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.043281 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: E0218 15:22:59.056164 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:59Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.059596 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.059633 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.059644 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.059663 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.059675 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: E0218 15:22:59.070869 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:22:59Z is after 2025-08-24T17:21:41Z" Feb 18 15:22:59 crc kubenswrapper[4968]: E0218 15:22:59.071043 4968 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.073021 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.073053 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.073066 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.073083 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.073100 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.176289 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.176327 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.176335 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.176358 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.176369 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.212948 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 17:02:09.143483327 +0000 UTC Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.230585 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.230719 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:22:59 crc kubenswrapper[4968]: E0218 15:22:59.230734 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:22:59 crc kubenswrapper[4968]: E0218 15:22:59.231430 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.278866 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.278909 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.278920 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.278938 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.278951 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.383877 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.383939 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.383957 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.383993 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.384019 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.487510 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.487582 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.487606 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.487642 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.487670 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.590896 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.590972 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.590997 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.591034 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.591061 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.694238 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.694311 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.694326 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.694348 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.694362 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.797535 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.797639 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.797705 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.797832 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.797855 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.901623 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.901670 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.901683 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.901701 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:22:59 crc kubenswrapper[4968]: I0218 15:22:59.901712 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:22:59Z","lastTransitionTime":"2026-02-18T15:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.004955 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.005007 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.005017 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.005035 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.005058 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:00Z","lastTransitionTime":"2026-02-18T15:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.108639 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.108718 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.108735 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.108842 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.108865 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:00Z","lastTransitionTime":"2026-02-18T15:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.211965 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.212028 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.212051 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.212084 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.212109 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:00Z","lastTransitionTime":"2026-02-18T15:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.213607 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 21:11:12.177782658 +0000 UTC Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.230033 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.230094 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:00 crc kubenswrapper[4968]: E0218 15:23:00.230198 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:00 crc kubenswrapper[4968]: E0218 15:23:00.230314 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.315544 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.315613 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.315631 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.315656 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.315674 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:00Z","lastTransitionTime":"2026-02-18T15:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.419392 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.419587 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.419615 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.419652 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.419677 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:00Z","lastTransitionTime":"2026-02-18T15:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.524127 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.524202 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.524228 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.524262 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.524284 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:00Z","lastTransitionTime":"2026-02-18T15:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.627656 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.627742 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.627813 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.627849 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.627876 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:00Z","lastTransitionTime":"2026-02-18T15:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.731085 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.731172 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.731205 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.731244 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.731269 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:00Z","lastTransitionTime":"2026-02-18T15:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.834822 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.835010 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.835032 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.835812 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.836069 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:00Z","lastTransitionTime":"2026-02-18T15:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.940715 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.940894 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.940925 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.940951 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:00 crc kubenswrapper[4968]: I0218 15:23:00.940964 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:00Z","lastTransitionTime":"2026-02-18T15:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.045014 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.045098 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.045114 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.045138 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.045153 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:01Z","lastTransitionTime":"2026-02-18T15:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.149040 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.149100 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.149112 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.149135 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.149149 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:01Z","lastTransitionTime":"2026-02-18T15:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.214597 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 13:48:26.480882425 +0000 UTC Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.230315 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.230338 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:01 crc kubenswrapper[4968]: E0218 15:23:01.230616 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:01 crc kubenswrapper[4968]: E0218 15:23:01.230830 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.250085 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.253113 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.253193 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.253221 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.253258 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.253284 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:01Z","lastTransitionTime":"2026-02-18T15:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.270489 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.305371 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.322770 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.345840 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:39Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0218 15:22:39.090619 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0218 15:22:39.090655 6647 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI0218 15:22:39.090680 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI0218 15:22:39.090693 6647 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:39.090705 6647 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:39.090602 6647 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0218 15:22:39.090903 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0218 15:22:39.090948 6647 factory.go:656] Stopping watch factory\\\\nI0218 15:22:39.090972 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0218 15:22:39.091376 6647 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0218 15:22:39.091558 6647 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0218 15:22:39.091665 6647 ovnkube.go:599] Stopped ovnkube\\\\nI0218 15:22:39.091737 6647 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 15:22:39.091917 6647 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.356809 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.357098 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.357191 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.357264 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.357339 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:01Z","lastTransitionTime":"2026-02-18T15:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.364594 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6ccc51a3ebdf422ba17225bac5935ff711e8ed81619d01a461692c3fb9083eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:51Z\\\",\\\"message\\\":\\\"2026-02-18T15:22:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae\\\\n2026-02-18T15:22:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae to /host/opt/cni/bin/\\\\n2026-02-18T15:22:06Z [verbose] multus-daemon started\\\\n2026-02-18T15:22:06Z [verbose] Readiness Indicator file check\\\\n2026-02-18T15:22:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.379957 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.395448 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca2cfa67-dc37-4b16-a924-0facaef81020\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b366a27a56c9622b6f0bba0cc98793876099f07cc7cd657ec8db748e3c79d4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8689dccaff6c24580b8defc45484e8b268076501a84b197ec61da7fa003680d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8689dccaff6c24580b8defc45484e8b268076501a84b197ec61da7fa003680d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.411906 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.426686 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.446362 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.460997 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.461039 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.461051 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.461069 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.461081 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:01Z","lastTransitionTime":"2026-02-18T15:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.466279 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.482934 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.499955 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.517428 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.531634 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.544465 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.563372 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.564283 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.564356 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.564369 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.564389 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.564400 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:01Z","lastTransitionTime":"2026-02-18T15:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.579790 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:01Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.668257 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.668319 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.668332 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.668352 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.668371 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:01Z","lastTransitionTime":"2026-02-18T15:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.771358 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.771403 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.771413 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.771429 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.771439 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:01Z","lastTransitionTime":"2026-02-18T15:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.874849 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.874915 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.874933 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.874953 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.874964 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:01Z","lastTransitionTime":"2026-02-18T15:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.978509 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.978567 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.978585 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.979918 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:01 crc kubenswrapper[4968]: I0218 15:23:01.979955 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:01Z","lastTransitionTime":"2026-02-18T15:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.082890 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.082945 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.082958 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.082979 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.082992 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:02Z","lastTransitionTime":"2026-02-18T15:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.185994 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.186064 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.186084 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.186113 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.186133 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:02Z","lastTransitionTime":"2026-02-18T15:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.215522 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 12:53:43.071783424 +0000 UTC Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.230057 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.230124 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:02 crc kubenswrapper[4968]: E0218 15:23:02.230265 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:02 crc kubenswrapper[4968]: E0218 15:23:02.230665 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.290550 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.290629 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.290652 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.290684 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.290706 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:02Z","lastTransitionTime":"2026-02-18T15:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.394703 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.394791 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.394815 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.394847 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.394869 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:02Z","lastTransitionTime":"2026-02-18T15:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.498190 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.498245 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.498259 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.498281 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.498293 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:02Z","lastTransitionTime":"2026-02-18T15:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.601434 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.601508 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.601527 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.601555 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.601575 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:02Z","lastTransitionTime":"2026-02-18T15:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.704347 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.704395 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.704410 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.704426 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.704435 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:02Z","lastTransitionTime":"2026-02-18T15:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.807875 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.807934 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.807948 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.807976 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.807988 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:02Z","lastTransitionTime":"2026-02-18T15:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.911415 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.911457 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.911467 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.911483 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:02 crc kubenswrapper[4968]: I0218 15:23:02.911494 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:02Z","lastTransitionTime":"2026-02-18T15:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.014475 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.014532 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.014546 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.014566 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.014584 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:03Z","lastTransitionTime":"2026-02-18T15:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.116861 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.116982 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.117009 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.117042 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.117067 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:03Z","lastTransitionTime":"2026-02-18T15:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.215835 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 18:20:20.706892097 +0000 UTC Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.220212 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.220253 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.220282 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.220301 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.220311 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:03Z","lastTransitionTime":"2026-02-18T15:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.229867 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.229899 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:03 crc kubenswrapper[4968]: E0218 15:23:03.230055 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:03 crc kubenswrapper[4968]: E0218 15:23:03.230265 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.324578 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.324840 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.324851 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.324868 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.324878 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:03Z","lastTransitionTime":"2026-02-18T15:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.427376 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.427425 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.427435 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.427455 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.427468 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:03Z","lastTransitionTime":"2026-02-18T15:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.530277 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.530337 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.530350 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.530377 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.530395 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:03Z","lastTransitionTime":"2026-02-18T15:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.634016 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.634078 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.634092 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.634111 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.634133 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:03Z","lastTransitionTime":"2026-02-18T15:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.737046 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.737099 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.737110 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.737128 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.737139 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:03Z","lastTransitionTime":"2026-02-18T15:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.839559 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.839618 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.839631 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.839653 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.839667 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:03Z","lastTransitionTime":"2026-02-18T15:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.942667 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.943028 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.943043 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.943063 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:03 crc kubenswrapper[4968]: I0218 15:23:03.943076 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:03Z","lastTransitionTime":"2026-02-18T15:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.046007 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.046074 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.046084 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.046109 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.046122 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:04Z","lastTransitionTime":"2026-02-18T15:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.149524 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.149598 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.149611 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.149681 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.149696 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:04Z","lastTransitionTime":"2026-02-18T15:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.216682 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 18:23:28.377425294 +0000 UTC Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.230025 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.230025 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:04 crc kubenswrapper[4968]: E0218 15:23:04.230183 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:04 crc kubenswrapper[4968]: E0218 15:23:04.230318 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.253013 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.253048 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.253060 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.253082 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.253098 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:04Z","lastTransitionTime":"2026-02-18T15:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.355845 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.355912 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.355930 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.355956 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.355977 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:04Z","lastTransitionTime":"2026-02-18T15:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.459436 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.459516 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.459536 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.459564 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.459589 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:04Z","lastTransitionTime":"2026-02-18T15:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.562133 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.562207 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.562218 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.562235 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.562247 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:04Z","lastTransitionTime":"2026-02-18T15:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.665541 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.665602 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.665613 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.665629 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.665639 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:04Z","lastTransitionTime":"2026-02-18T15:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.768369 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.768430 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.768446 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.768468 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.768484 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:04Z","lastTransitionTime":"2026-02-18T15:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.871128 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.871194 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.871219 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.871254 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.871274 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:04Z","lastTransitionTime":"2026-02-18T15:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.974678 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.974722 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.974732 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.974771 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:04 crc kubenswrapper[4968]: I0218 15:23:04.974786 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:04Z","lastTransitionTime":"2026-02-18T15:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.077341 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.077386 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.077400 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.077420 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.077432 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:05Z","lastTransitionTime":"2026-02-18T15:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.152354 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.152657 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:09.152612551 +0000 UTC m=+148.538057543 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.180416 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.180469 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.180482 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.180503 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.180518 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:05Z","lastTransitionTime":"2026-02-18T15:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.217261 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 20:35:22.771060756 +0000 UTC Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.230157 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.230157 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.230547 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.230867 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.253975 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.254025 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.254053 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.254103 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.254161 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.254190 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.254207 4968 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.255158 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 15:24:09.255108846 +0000 UTC m=+148.640553708 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.255282 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.255342 4968 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.255368 4968 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.255415 4968 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.255288 4968 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.255505 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 15:24:09.255443006 +0000 UTC m=+148.640887868 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.255737 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:24:09.255624151 +0000 UTC m=+148.641069013 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 15:23:05 crc kubenswrapper[4968]: E0218 15:23:05.255784 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 15:24:09.255771745 +0000 UTC m=+148.641216607 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.283281 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.283334 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.283354 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.283383 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.283403 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:05Z","lastTransitionTime":"2026-02-18T15:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.386653 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.386705 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.386716 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.386737 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.386775 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:05Z","lastTransitionTime":"2026-02-18T15:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.490413 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.490457 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.490468 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.490488 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.490501 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:05Z","lastTransitionTime":"2026-02-18T15:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.593669 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.593724 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.593738 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.593780 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.593796 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:05Z","lastTransitionTime":"2026-02-18T15:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.696663 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.696729 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.696804 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.696833 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.696856 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:05Z","lastTransitionTime":"2026-02-18T15:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.799897 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.799946 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.799962 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.799984 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.799998 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:05Z","lastTransitionTime":"2026-02-18T15:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.904059 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.904129 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.904147 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.904177 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:05 crc kubenswrapper[4968]: I0218 15:23:05.904196 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:05Z","lastTransitionTime":"2026-02-18T15:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.007172 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.007234 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.007252 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.007279 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.007301 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:06Z","lastTransitionTime":"2026-02-18T15:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.111411 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.111489 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.111515 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.111548 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.111572 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:06Z","lastTransitionTime":"2026-02-18T15:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.215296 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.215359 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.215371 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.215398 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.215412 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:06Z","lastTransitionTime":"2026-02-18T15:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.218277 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 14:39:40.678555307 +0000 UTC Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.230697 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.230734 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:06 crc kubenswrapper[4968]: E0218 15:23:06.230911 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:06 crc kubenswrapper[4968]: E0218 15:23:06.231086 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.321325 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.321659 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.321793 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.321951 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.322053 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:06Z","lastTransitionTime":"2026-02-18T15:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.425799 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.425884 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.425910 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.425949 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.425973 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:06Z","lastTransitionTime":"2026-02-18T15:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.528635 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.528685 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.528695 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.528717 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.528729 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:06Z","lastTransitionTime":"2026-02-18T15:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.631286 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.631338 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.631350 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.631371 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.631385 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:06Z","lastTransitionTime":"2026-02-18T15:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.734088 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.734413 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.734431 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.734448 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.734458 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:06Z","lastTransitionTime":"2026-02-18T15:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.836678 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.836723 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.836733 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.836770 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.836787 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:06Z","lastTransitionTime":"2026-02-18T15:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.939438 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.939726 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.939828 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.939913 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:06 crc kubenswrapper[4968]: I0218 15:23:06.940037 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:06Z","lastTransitionTime":"2026-02-18T15:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.043042 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.043096 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.043114 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.043142 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.043164 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:07Z","lastTransitionTime":"2026-02-18T15:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.146134 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.146184 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.146200 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.146220 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.146236 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:07Z","lastTransitionTime":"2026-02-18T15:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.218896 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 17:38:15.542031041 +0000 UTC Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.230456 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.230521 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:07 crc kubenswrapper[4968]: E0218 15:23:07.230897 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:07 crc kubenswrapper[4968]: E0218 15:23:07.231005 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.231166 4968 scope.go:117] "RemoveContainer" containerID="ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.249517 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.249561 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.249578 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.249604 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.249622 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:07Z","lastTransitionTime":"2026-02-18T15:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.353708 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.354426 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.354449 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.354480 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.354502 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:07Z","lastTransitionTime":"2026-02-18T15:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.456702 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.456768 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.456784 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.456806 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.456822 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:07Z","lastTransitionTime":"2026-02-18T15:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.560234 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.560295 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.560308 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.560324 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.560336 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:07Z","lastTransitionTime":"2026-02-18T15:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.663724 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.663830 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.663848 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.663871 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.663892 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:07Z","lastTransitionTime":"2026-02-18T15:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.767037 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.767080 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.767089 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.767105 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.767115 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:07Z","lastTransitionTime":"2026-02-18T15:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.781472 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/2.log" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.785199 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3"} Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.785720 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.806979 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.823911 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.845324 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.866496 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.870404 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.870440 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.870449 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.870465 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.870479 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:07Z","lastTransitionTime":"2026-02-18T15:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.881091 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.896869 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.914835 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.933616 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.948512 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.963089 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.974118 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.974177 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.974193 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.974235 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.974254 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:07Z","lastTransitionTime":"2026-02-18T15:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.980221 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:07 crc kubenswrapper[4968]: I0218 15:23:07.993790 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:07Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.005575 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.017415 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca2cfa67-dc37-4b16-a924-0facaef81020\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b366a27a56c9622b6f0bba0cc98793876099f07cc7cd657ec8db748e3c79d4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8689dccaff6c24580b8defc45484e8b268076501a84b197ec61da7fa003680d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8689dccaff6c24580b8defc45484e8b268076501a84b197ec61da7fa003680d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.038082 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.053045 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.073693 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:39Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0218 15:22:39.090619 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0218 15:22:39.090655 6647 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI0218 15:22:39.090680 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI0218 15:22:39.090693 6647 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:39.090705 6647 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:39.090602 6647 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0218 15:22:39.090903 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0218 15:22:39.090948 6647 factory.go:656] Stopping watch factory\\\\nI0218 15:22:39.090972 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0218 15:22:39.091376 6647 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0218 15:22:39.091558 6647 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0218 15:22:39.091665 6647 ovnkube.go:599] Stopped ovnkube\\\\nI0218 15:22:39.091737 6647 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 15:22:39.091917 6647 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:23:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.077872 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.077929 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.077939 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.077958 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.077969 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:08Z","lastTransitionTime":"2026-02-18T15:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.090646 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6ccc51a3ebdf422ba17225bac5935ff711e8ed81619d01a461692c3fb9083eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:51Z\\\",\\\"message\\\":\\\"2026-02-18T15:22:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae\\\\n2026-02-18T15:22:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae to /host/opt/cni/bin/\\\\n2026-02-18T15:22:06Z [verbose] multus-daemon started\\\\n2026-02-18T15:22:06Z [verbose] Readiness Indicator file check\\\\n2026-02-18T15:22:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.104510 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.181593 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.181635 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.181645 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.181662 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.181674 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:08Z","lastTransitionTime":"2026-02-18T15:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.219578 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 16:38:38.849868244 +0000 UTC Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.231051 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.231172 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:08 crc kubenswrapper[4968]: E0218 15:23:08.231281 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:08 crc kubenswrapper[4968]: E0218 15:23:08.231383 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.285155 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.285204 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.285216 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.285234 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.285245 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:08Z","lastTransitionTime":"2026-02-18T15:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.388085 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.388126 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.388137 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.388154 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.388166 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:08Z","lastTransitionTime":"2026-02-18T15:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.490618 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.490665 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.490677 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.490703 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.490716 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:08Z","lastTransitionTime":"2026-02-18T15:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.593790 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.593836 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.593849 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.593871 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.593885 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:08Z","lastTransitionTime":"2026-02-18T15:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.696141 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.696181 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.696192 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.696208 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.696218 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:08Z","lastTransitionTime":"2026-02-18T15:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.790361 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/3.log" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.791235 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/2.log" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.794041 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" exitCode=1 Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.794084 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3"} Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.794155 4968 scope.go:117] "RemoveContainer" containerID="ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.795108 4968 scope.go:117] "RemoveContainer" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" Feb 18 15:23:08 crc kubenswrapper[4968]: E0218 15:23:08.795343 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.802257 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.802285 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.802295 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.802313 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.802325 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:08Z","lastTransitionTime":"2026-02-18T15:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.815828 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca2cfa67-dc37-4b16-a924-0facaef81020\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b366a27a56c9622b6f0bba0cc98793876099f07cc7cd657ec8db748e3c79d4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8689dccaff6c24580b8defc45484e8b268076501a84b197ec61da7fa003680d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8689dccaff6c24580b8defc45484e8b268076501a84b197ec61da7fa003680d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.853086 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.868384 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.889313 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffd1edc9fd4a6c867238de4d0afa7c0ca19a2ab4d082536aca5aa1d2baf7213a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:39Z\\\",\\\"message\\\":\\\"andler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0218 15:22:39.090619 6647 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0218 15:22:39.090655 6647 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0218 15:22:39.090668 6647 handler.go:208] Removed *v1.Node event handler 2\\\\nI0218 15:22:39.090680 6647 handler.go:208] Removed *v1.Node event handler 7\\\\nI0218 15:22:39.090693 6647 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0218 15:22:39.090705 6647 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0218 15:22:39.090602 6647 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0218 15:22:39.090903 6647 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0218 15:22:39.090948 6647 factory.go:656] Stopping watch factory\\\\nI0218 15:22:39.090972 6647 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0218 15:22:39.091376 6647 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0218 15:22:39.091558 6647 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0218 15:22:39.091665 6647 ovnkube.go:599] Stopped ovnkube\\\\nI0218 15:22:39.091737 6647 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 15:22:39.091917 6647 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:23:08Z\\\",\\\"message\\\":\\\"p[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 15:23:08.089077 7073 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 15:23:08.090306 7073 ovnkube.go:599] Stopped ovnkube\\\\nI0218 15:23:08.090364 7073 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 15:23:08.090468 7073 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:23:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.905033 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.905224 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.905242 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.905262 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.905275 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:08Z","lastTransitionTime":"2026-02-18T15:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.906101 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6ccc51a3ebdf422ba17225bac5935ff711e8ed81619d01a461692c3fb9083eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:51Z\\\",\\\"message\\\":\\\"2026-02-18T15:22:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae\\\\n2026-02-18T15:22:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae to /host/opt/cni/bin/\\\\n2026-02-18T15:22:06Z [verbose] multus-daemon started\\\\n2026-02-18T15:22:06Z [verbose] Readiness Indicator file check\\\\n2026-02-18T15:22:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.918029 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.931932 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.942866 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.954411 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.969112 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:08 crc kubenswrapper[4968]: I0218 15:23:08.992789 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:08Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.005053 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.008034 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.008327 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.008362 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.008407 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.008419 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.018995 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.031816 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.045901 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.055590 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.069999 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.080644 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.091706 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.112036 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.112081 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.112096 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.112118 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.112131 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.215912 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.215965 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.215985 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.216014 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.216028 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.220127 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 13:36:36.325985128 +0000 UTC Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.222608 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.222646 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.222657 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.222676 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.222688 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.231030 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:09 crc kubenswrapper[4968]: E0218 15:23:09.231192 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.231506 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:09 crc kubenswrapper[4968]: E0218 15:23:09.231583 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:09 crc kubenswrapper[4968]: E0218 15:23:09.239425 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.244284 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.244328 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.244341 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.244359 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.244372 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: E0218 15:23:09.260479 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.265331 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.265413 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.265437 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.265467 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.265485 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: E0218 15:23:09.284344 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.291109 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.291150 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.291166 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.291187 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.291200 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: E0218 15:23:09.308997 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.313909 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.313954 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.313967 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.313987 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.313998 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: E0218 15:23:09.332970 4968 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T15:23:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0e3deaf-b79c-4e72-8a61-06dcf6ae6c27\\\",\\\"systemUUID\\\":\\\"9e33db9e-c77c-46df-b8f6-fcfb068b9f9d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: E0218 15:23:09.333104 4968 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.334819 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.334870 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.334881 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.334897 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.334908 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.437644 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.437692 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.437702 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.437721 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.437732 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.541269 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.541316 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.541329 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.541355 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.541369 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.644139 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.644200 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.644212 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.644231 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.644245 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.750371 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.750439 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.750454 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.750801 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.750824 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.801316 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/3.log" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.805310 4968 scope.go:117] "RemoveContainer" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" Feb 18 15:23:09 crc kubenswrapper[4968]: E0218 15:23:09.805541 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.821914 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.839287 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.853592 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.856076 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.857124 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.857156 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.857179 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.857194 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.868637 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.881455 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.893156 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.906022 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.919058 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.933457 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.945369 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca2cfa67-dc37-4b16-a924-0facaef81020\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b366a27a56c9622b6f0bba0cc98793876099f07cc7cd657ec8db748e3c79d4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8689dccaff6c24580b8defc45484e8b268076501a84b197ec61da7fa003680d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8689dccaff6c24580b8defc45484e8b268076501a84b197ec61da7fa003680d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.959973 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.960021 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.960051 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.960072 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.960086 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:09Z","lastTransitionTime":"2026-02-18T15:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.966100 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.978446 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:09 crc kubenswrapper[4968]: I0218 15:23:09.997895 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:23:08Z\\\",\\\"message\\\":\\\"p[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 15:23:08.089077 7073 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 15:23:08.090306 7073 ovnkube.go:599] Stopped ovnkube\\\\nI0218 15:23:08.090364 7073 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 15:23:08.090468 7073 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:23:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:09Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.010841 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6ccc51a3ebdf422ba17225bac5935ff711e8ed81619d01a461692c3fb9083eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:51Z\\\",\\\"message\\\":\\\"2026-02-18T15:22:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae\\\\n2026-02-18T15:22:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae to /host/opt/cni/bin/\\\\n2026-02-18T15:22:06Z [verbose] multus-daemon started\\\\n2026-02-18T15:22:06Z [verbose] Readiness Indicator file check\\\\n2026-02-18T15:22:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.021296 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.035541 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.049392 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.062215 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.063245 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.063313 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.063327 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.063345 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.063372 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:10Z","lastTransitionTime":"2026-02-18T15:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.079364 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:10Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.166326 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.166368 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.166408 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.166428 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.166441 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:10Z","lastTransitionTime":"2026-02-18T15:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.220739 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 21:31:24.297804518 +0000 UTC Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.230306 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.230307 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:10 crc kubenswrapper[4968]: E0218 15:23:10.230537 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:10 crc kubenswrapper[4968]: E0218 15:23:10.230662 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.269977 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.270049 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.270063 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.270086 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.270098 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:10Z","lastTransitionTime":"2026-02-18T15:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.373131 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.373835 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.373935 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.373989 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.374003 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:10Z","lastTransitionTime":"2026-02-18T15:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.477067 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.477118 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.477127 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.477142 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.477153 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:10Z","lastTransitionTime":"2026-02-18T15:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.580696 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.580779 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.580801 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.580831 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.580849 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:10Z","lastTransitionTime":"2026-02-18T15:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.683452 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.683508 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.683521 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.683545 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.683558 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:10Z","lastTransitionTime":"2026-02-18T15:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.786937 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.787004 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.787026 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.787049 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.787063 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:10Z","lastTransitionTime":"2026-02-18T15:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.890239 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.890332 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.890353 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.890385 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.890414 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:10Z","lastTransitionTime":"2026-02-18T15:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.994350 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.994426 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.994449 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.994479 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:10 crc kubenswrapper[4968]: I0218 15:23:10.994503 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:10Z","lastTransitionTime":"2026-02-18T15:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.097689 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.097792 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.097808 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.097827 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.097865 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:11Z","lastTransitionTime":"2026-02-18T15:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.201389 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.201481 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.201502 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.201531 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.201557 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:11Z","lastTransitionTime":"2026-02-18T15:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.221976 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 20:06:00.288538091 +0000 UTC Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.230781 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.230855 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:11 crc kubenswrapper[4968]: E0218 15:23:11.230950 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:11 crc kubenswrapper[4968]: E0218 15:23:11.231146 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.258842 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"435c6f94-b91e-4ce0-8407-5227f3a5078f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:23:08Z\\\",\\\"message\\\":\\\"p[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 15:23:08.089077 7073 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 15:23:08.090306 7073 ovnkube.go:599] Stopped ovnkube\\\\nI0218 15:23:08.090364 7073 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 15:23:08.090468 7073 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:23:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5hmgv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z2jkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.277349 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2qq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d01781a-6a78-49a2-80c7-9ac02c810e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6ccc51a3ebdf422ba17225bac5935ff711e8ed81619d01a461692c3fb9083eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T15:22:51Z\\\",\\\"message\\\":\\\"2026-02-18T15:22:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae\\\\n2026-02-18T15:22:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6d2bf6f-46e5-4227-846f-83065d8d92ae to /host/opt/cni/bin/\\\\n2026-02-18T15:22:06Z [verbose] multus-daemon started\\\\n2026-02-18T15:22:06Z [verbose] Readiness Indicator file check\\\\n2026-02-18T15:22:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsdhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2qq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.294181 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9185b44b-bac7-458b-b4d9-4c389da07c14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5dcz8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7sk6k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.304256 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.304317 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.304336 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.304363 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.304382 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:11Z","lastTransitionTime":"2026-02-18T15:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.309651 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca2cfa67-dc37-4b16-a924-0facaef81020\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b366a27a56c9622b6f0bba0cc98793876099f07cc7cd657ec8db748e3c79d4ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8689dccaff6c24580b8defc45484e8b268076501a84b197ec61da7fa003680d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8689dccaff6c24580b8defc45484e8b268076501a84b197ec61da7fa003680d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.343789 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe2022e6-63b3-4415-a06b-f4b76ae4b3ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ada7830aeb7c404dcacff5b2ba9dc11733eda783ebeda0362ae6a223413d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74bfc561fe73d5d63790f4b8b19b71da4fa9daf71f78d2685d9ecf523b345bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d662284b75cd37846b700fddfdbab44dd11dd971e5c850a8561e832167c00aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59a7b0c444380d931ac715f896dde4a7b44bc883029d62aaa3c633538b4dd19f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55c83987a9b07126dc9a43c1b688450b7eb3a6f9a8d486786e3dab56ccaa2b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff1dd8cf4591f986e40f9903eb74c0573e18056feec2a51fed04ded34bb6183\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cf1964636d0a0addbbb609730e53e4a0e666db411756b1274af4cbc20a6313\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8243d3b6e275bee7f2263ebdaee18ea7f016b173277e9501313ce33a1f40ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.359151 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be66c6b93a1e9941d06c98437b68fb50767201c50221f10892ca68d03ba3ffdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.373793 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6694896c-93a1-47ac-a079-20501cf9909e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24f168a84149f258e7a1cd827ca761e85c45dd6571e89783b5b0005f08f6630a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://074a59aa0527c2cc6e300d4b64bd42a99538029ba693715ef12a85e5ac5c92bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3f36e47f4e60180aad9a1dc1192ef70cc5b6ec2b8ce107171c7a737745091d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ab929174461745f0649d708fa4777776151dd8202ac7da3de8fde2f275fc18d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00a6ec08f1cea86ce5a154215cc94a9231af854ee43324d23bfa33dae700ae17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a72ff2c40caffcbabe4c93a623e2a40b359d752235217096e12b59146730cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbab067cfc70716f97ea497932327b96a3bb11c89c8243e25b02a9a52cf2c4c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-59cjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5rzpj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.391836 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c34049-85fe-4d7b-af06-64a247724267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T15:21:55Z\\\",\\\"message\\\":\\\"W0218 15:21:44.648109 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 15:21:44.649028 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771428104 cert, and key in /tmp/serving-cert-172314993/serving-signer.crt, /tmp/serving-cert-172314993/serving-signer.key\\\\nI0218 15:21:45.013693 1 observer_polling.go:159] Starting file observer\\\\nW0218 15:21:45.016942 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 15:21:45.017268 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 15:21:45.018261 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-172314993/tls.crt::/tmp/serving-cert-172314993/tls.key\\\\\\\"\\\\nF0218 15:21:55.348239 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.407704 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.407783 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.407803 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.407831 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.407853 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:11Z","lastTransitionTime":"2026-02-18T15:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.408580 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.424440 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.442486 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aae5eebcd41f40e44dc1a3cbf90ef8923cdc277e75afe85b22152eb7b86a9b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdc28f0a5f7afb2846afc88a00f17c5266885f2e350dd5dcb0ed8b83af315e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.456386 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390179a-7a57-4696-ab4c-a0c91eeabea2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ecc91696cb2881ea0c33427d07a9623043f5e00ab81e25241fe38f00620aad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhzqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.467218 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-78lnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"358a1641-853a-4075-8f59-de12eb00b601\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ac4ffeec207ad4fa8369e384334e707364263676ceeb5ad83d6c41d00de9515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4wsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-78lnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.481724 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd365caa-2db2-407a-b264-d18a0d06fe06\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e445f58aecf32ecadbf665d74975d525b730b67fa01c0e0578d190c397ab632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b90a2a717f79e0d02cefc2df574203ac37e8eaa65453036b00f4840ae4f1b476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qd8r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qsjc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.497736 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69845080-bcf6-4e9c-a4dd-d3df41cd7b98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6d4444b1502c0112ac6f91a591c565e04dd82a0fc93fd3f5da903e49bd7a057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3597bd758f398de98eaf0e040759c3352086ee8b1632c8fb11994981b2a8272\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d8613519cdbfb0f43ff826af5ae2f8c9af8baff39a1b5a2ebe4d85f31e0c4b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.511009 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.511047 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.511059 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.511080 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.511095 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:11Z","lastTransitionTime":"2026-02-18T15:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.517074 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddfbc3c0fce7c3c4f08d531ed304c6284f263e801cfc4a334524d09d41c1a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.538417 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.553193 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36cc0c32-cccc-494e-b1ea-6226e7e880ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:21:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e530883430d9220650a000aff99d358093aac53d7d27a5c72dcb345d498477\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd8a7fa47744269afead6e2bb1ee8051036d5285197aedd5822efe175e5a36ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06d3c2293997753e1450f64243c1b40c1e93cb76f045250d86ccbe669729dd6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:21:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff5c809a1b5177c23e3ad9be6af26ede88f2a0a01dc87704b7a36f513065f2a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T15:21:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T15:21:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:21:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.568956 4968 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9bae90c-908f-40fd-8373-4bf7f9aaede6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T15:22:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://959bf219cf428b24856917c4b74f288c49bcfb2f434e09b4db90b5919d7bf12f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T15:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qrwt5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T15:22:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xnhwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T15:23:11Z is after 2025-08-24T17:21:41Z" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.614854 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.614935 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.614957 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.614986 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.615005 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:11Z","lastTransitionTime":"2026-02-18T15:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.717959 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.718017 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.718032 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.718054 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.718074 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:11Z","lastTransitionTime":"2026-02-18T15:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.821628 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.821680 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.821694 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.821713 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.821730 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:11Z","lastTransitionTime":"2026-02-18T15:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.925924 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.926005 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.926025 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.926055 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:11 crc kubenswrapper[4968]: I0218 15:23:11.926074 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:11Z","lastTransitionTime":"2026-02-18T15:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.029347 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.029419 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.029455 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.029484 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.029504 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:12Z","lastTransitionTime":"2026-02-18T15:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.132593 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.132672 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.132702 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.132735 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.132789 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:12Z","lastTransitionTime":"2026-02-18T15:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.222315 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 23:25:15.208202442 +0000 UTC Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.229725 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.229736 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:12 crc kubenswrapper[4968]: E0218 15:23:12.230042 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:12 crc kubenswrapper[4968]: E0218 15:23:12.230099 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.236544 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.236592 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.236608 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.236631 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.236648 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:12Z","lastTransitionTime":"2026-02-18T15:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.349409 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.349476 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.349496 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.349524 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.349542 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:12Z","lastTransitionTime":"2026-02-18T15:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.453803 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.453905 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.453931 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.453961 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.453982 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:12Z","lastTransitionTime":"2026-02-18T15:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.557275 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.557326 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.557336 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.557356 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.557369 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:12Z","lastTransitionTime":"2026-02-18T15:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.660297 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.660352 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.660378 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.660405 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.660423 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:12Z","lastTransitionTime":"2026-02-18T15:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.764496 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.764556 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.764574 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.764599 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.764617 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:12Z","lastTransitionTime":"2026-02-18T15:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.867248 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.867304 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.867316 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.867337 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.867353 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:12Z","lastTransitionTime":"2026-02-18T15:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.970234 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.970294 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.970305 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.970326 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:12 crc kubenswrapper[4968]: I0218 15:23:12.970338 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:12Z","lastTransitionTime":"2026-02-18T15:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.073832 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.073897 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.073912 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.073937 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.073954 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:13Z","lastTransitionTime":"2026-02-18T15:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.178007 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.178058 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.178071 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.178091 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.178104 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:13Z","lastTransitionTime":"2026-02-18T15:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.223327 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 16:51:11.836377724 +0000 UTC Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.229839 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.230036 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:13 crc kubenswrapper[4968]: E0218 15:23:13.230313 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:13 crc kubenswrapper[4968]: E0218 15:23:13.230611 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.281435 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.281532 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.281543 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.281583 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.281599 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:13Z","lastTransitionTime":"2026-02-18T15:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.385424 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.385468 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.385482 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.385502 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.385512 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:13Z","lastTransitionTime":"2026-02-18T15:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.489013 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.489085 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.489099 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.489121 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.489135 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:13Z","lastTransitionTime":"2026-02-18T15:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.592550 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.592609 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.592623 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.592642 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.592658 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:13Z","lastTransitionTime":"2026-02-18T15:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.696543 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.696613 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.696633 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.696662 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.696688 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:13Z","lastTransitionTime":"2026-02-18T15:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.800683 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.800827 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.800854 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.800891 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.800913 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:13Z","lastTransitionTime":"2026-02-18T15:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.904787 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.904852 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.904871 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.904897 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:13 crc kubenswrapper[4968]: I0218 15:23:13.904915 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:13Z","lastTransitionTime":"2026-02-18T15:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.008288 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.008341 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.008354 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.008374 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.008386 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:14Z","lastTransitionTime":"2026-02-18T15:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.112029 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.112086 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.112100 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.112124 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.112141 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:14Z","lastTransitionTime":"2026-02-18T15:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.215489 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.215543 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.215552 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.215568 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.215579 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:14Z","lastTransitionTime":"2026-02-18T15:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.223784 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 11:25:58.992736866 +0000 UTC Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.230383 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.230450 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:14 crc kubenswrapper[4968]: E0218 15:23:14.231077 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:14 crc kubenswrapper[4968]: E0218 15:23:14.231227 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.318043 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.318122 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.318142 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.318173 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.318194 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:14Z","lastTransitionTime":"2026-02-18T15:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.422331 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.422485 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.422508 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.422538 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.422561 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:14Z","lastTransitionTime":"2026-02-18T15:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.526672 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.526773 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.526789 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.526812 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.526834 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:14Z","lastTransitionTime":"2026-02-18T15:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.630227 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.630329 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.630354 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.630388 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.630555 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:14Z","lastTransitionTime":"2026-02-18T15:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.733272 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.733331 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.733348 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.733376 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.733394 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:14Z","lastTransitionTime":"2026-02-18T15:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.837022 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.837085 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.837104 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.837135 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.837154 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:14Z","lastTransitionTime":"2026-02-18T15:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.940970 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.941077 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.941091 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.941111 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:14 crc kubenswrapper[4968]: I0218 15:23:14.941481 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:14Z","lastTransitionTime":"2026-02-18T15:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.045089 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.045198 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.045238 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.045328 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.046041 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:15Z","lastTransitionTime":"2026-02-18T15:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.148965 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.149039 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.149063 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.149095 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.149133 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:15Z","lastTransitionTime":"2026-02-18T15:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.224403 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 19:32:35.444790425 +0000 UTC Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.230023 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.230023 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:15 crc kubenswrapper[4968]: E0218 15:23:15.230322 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:15 crc kubenswrapper[4968]: E0218 15:23:15.230398 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.252391 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.252472 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.252499 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.252534 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.252564 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:15Z","lastTransitionTime":"2026-02-18T15:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.355847 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.355925 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.355950 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.355983 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.356010 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:15Z","lastTransitionTime":"2026-02-18T15:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.459529 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.459588 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.459608 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.459633 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.459652 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:15Z","lastTransitionTime":"2026-02-18T15:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.563857 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.563917 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.563934 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.563961 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.563977 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:15Z","lastTransitionTime":"2026-02-18T15:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.666692 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.666731 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.666742 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.666772 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.666783 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:15Z","lastTransitionTime":"2026-02-18T15:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.770860 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.770912 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.770921 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.770938 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.770948 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:15Z","lastTransitionTime":"2026-02-18T15:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.874800 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.874857 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.874870 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.874892 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.874905 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:15Z","lastTransitionTime":"2026-02-18T15:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.978574 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.978642 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.978655 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.978678 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:15 crc kubenswrapper[4968]: I0218 15:23:15.978693 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:15Z","lastTransitionTime":"2026-02-18T15:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.081995 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.082055 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.082078 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.082106 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.082125 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:16Z","lastTransitionTime":"2026-02-18T15:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.185564 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.185633 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.185657 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.185692 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.185717 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:16Z","lastTransitionTime":"2026-02-18T15:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.225356 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 06:31:18.228946999 +0000 UTC Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.229588 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.229644 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:16 crc kubenswrapper[4968]: E0218 15:23:16.229769 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:16 crc kubenswrapper[4968]: E0218 15:23:16.230052 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.289551 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.289616 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.289635 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.289661 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.289681 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:16Z","lastTransitionTime":"2026-02-18T15:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.393414 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.393472 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.393491 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.393519 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.393540 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:16Z","lastTransitionTime":"2026-02-18T15:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.497656 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.497733 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.497796 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.497832 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.497856 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:16Z","lastTransitionTime":"2026-02-18T15:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.601101 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.601166 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.601176 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.601194 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.601205 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:16Z","lastTransitionTime":"2026-02-18T15:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.704615 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.704683 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.704703 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.704733 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.704782 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:16Z","lastTransitionTime":"2026-02-18T15:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.808142 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.808205 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.808218 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.808241 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.808255 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:16Z","lastTransitionTime":"2026-02-18T15:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.910612 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.910663 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.910677 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.910702 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:16 crc kubenswrapper[4968]: I0218 15:23:16.910717 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:16Z","lastTransitionTime":"2026-02-18T15:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.013157 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.013207 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.013221 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.013240 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.013251 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:17Z","lastTransitionTime":"2026-02-18T15:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.116973 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.117029 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.117050 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.117068 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.117078 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:17Z","lastTransitionTime":"2026-02-18T15:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.220451 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.220506 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.220515 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.220531 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.220542 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:17Z","lastTransitionTime":"2026-02-18T15:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.225738 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 08:46:38.606757664 +0000 UTC Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.230271 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.230359 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:17 crc kubenswrapper[4968]: E0218 15:23:17.230483 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:17 crc kubenswrapper[4968]: E0218 15:23:17.230601 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.323270 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.323324 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.323337 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.323357 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.323374 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:17Z","lastTransitionTime":"2026-02-18T15:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.425897 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.425947 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.425959 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.425975 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.425987 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:17Z","lastTransitionTime":"2026-02-18T15:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.529491 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.529547 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.529566 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.529588 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.529599 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:17Z","lastTransitionTime":"2026-02-18T15:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.632696 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.632761 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.632772 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.632793 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.632805 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:17Z","lastTransitionTime":"2026-02-18T15:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.735796 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.735852 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.735863 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.735886 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.735896 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:17Z","lastTransitionTime":"2026-02-18T15:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.837865 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.837915 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.837927 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.837946 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.837960 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:17Z","lastTransitionTime":"2026-02-18T15:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.941976 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.942042 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.942054 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.942074 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:17 crc kubenswrapper[4968]: I0218 15:23:17.942085 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:17Z","lastTransitionTime":"2026-02-18T15:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.045585 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.045647 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.045667 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.045688 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.045710 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:18Z","lastTransitionTime":"2026-02-18T15:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.149239 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.149313 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.149333 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.149363 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.149384 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:18Z","lastTransitionTime":"2026-02-18T15:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.228224 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 22:13:52.654453903 +0000 UTC Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.229809 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.229980 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:18 crc kubenswrapper[4968]: E0218 15:23:18.230204 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:18 crc kubenswrapper[4968]: E0218 15:23:18.230322 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.252553 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.252973 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.253061 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.253157 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.253230 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:18Z","lastTransitionTime":"2026-02-18T15:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.357858 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.358378 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.358605 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.358874 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.359136 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:18Z","lastTransitionTime":"2026-02-18T15:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.463668 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.463734 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.463773 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.463799 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.463816 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:18Z","lastTransitionTime":"2026-02-18T15:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.567418 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.567477 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.567492 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.567518 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.567533 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:18Z","lastTransitionTime":"2026-02-18T15:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.670403 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.670467 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.670480 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.670502 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.670519 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:18Z","lastTransitionTime":"2026-02-18T15:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.773723 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.773813 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.773832 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.773858 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.773879 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:18Z","lastTransitionTime":"2026-02-18T15:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.876685 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.876776 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.876793 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.876816 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.876829 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:18Z","lastTransitionTime":"2026-02-18T15:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.980102 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.980190 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.980215 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.980282 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:18 crc kubenswrapper[4968]: I0218 15:23:18.980308 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:18Z","lastTransitionTime":"2026-02-18T15:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.083936 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.083981 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.083990 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.084005 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.084015 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:19Z","lastTransitionTime":"2026-02-18T15:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.187327 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.187405 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.187426 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.187456 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.187473 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:19Z","lastTransitionTime":"2026-02-18T15:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.228937 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 05:00:21.006988412 +0000 UTC Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.230284 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.230332 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:19 crc kubenswrapper[4968]: E0218 15:23:19.230549 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:19 crc kubenswrapper[4968]: E0218 15:23:19.230710 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.291533 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.291612 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.291639 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.291675 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.291699 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:19Z","lastTransitionTime":"2026-02-18T15:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.395183 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.395242 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.395254 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.395275 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.395290 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:19Z","lastTransitionTime":"2026-02-18T15:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.498086 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.498138 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.498150 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.498171 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.498183 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:19Z","lastTransitionTime":"2026-02-18T15:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.566561 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.566619 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.566634 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.566653 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.566665 4968 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T15:23:19Z","lastTransitionTime":"2026-02-18T15:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.621503 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d"] Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.621992 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.624026 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.625735 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.625779 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.626159 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.638038 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:19 crc kubenswrapper[4968]: E0218 15:23:19.638204 4968 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:23:19 crc kubenswrapper[4968]: E0218 15:23:19.638269 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs podName:9185b44b-bac7-458b-b4d9-4c389da07c14 nodeName:}" failed. No retries permitted until 2026-02-18 15:24:23.638250982 +0000 UTC m=+163.023695844 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs") pod "network-metrics-daemon-7sk6k" (UID: "9185b44b-bac7-458b-b4d9-4c389da07c14") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.693972 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-m2qq8" podStartSLOduration=78.693943134 podStartE2EDuration="1m18.693943134s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:23:19.693238484 +0000 UTC m=+99.078683356" watchObservedRunningTime="2026-02-18 15:23:19.693943134 +0000 UTC m=+99.079388006" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.718193 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=24.71817033 podStartE2EDuration="24.71817033s" podCreationTimestamp="2026-02-18 15:22:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:23:19.717414428 +0000 UTC m=+99.102859300" watchObservedRunningTime="2026-02-18 15:23:19.71817033 +0000 UTC m=+99.103615192" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.739415 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.739516 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.739558 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.739635 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.739676 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.754060 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=75.754023614 podStartE2EDuration="1m15.754023614s" podCreationTimestamp="2026-02-18 15:22:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:23:19.751253543 +0000 UTC m=+99.136698415" watchObservedRunningTime="2026-02-18 15:23:19.754023614 +0000 UTC m=+99.139468496" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.785886 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-5rzpj" podStartSLOduration=78.785855651 podStartE2EDuration="1m18.785855651s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:23:19.785559182 +0000 UTC m=+99.171004084" watchObservedRunningTime="2026-02-18 15:23:19.785855651 +0000 UTC m=+99.171300523" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.835475 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.835450245 podStartE2EDuration="1m18.835450245s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:23:19.821985673 +0000 UTC m=+99.207430535" watchObservedRunningTime="2026-02-18 15:23:19.835450245 +0000 UTC m=+99.220895107" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.840842 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.840901 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.840941 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.840953 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.840980 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.841043 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.841117 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.842074 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.848850 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.859559 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a49d83c1-4f95-4c5b-acd6-e4836ad06be3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qq76d\" (UID: \"a49d83c1-4f95-4c5b-acd6-e4836ad06be3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.899080 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-t5rmj" podStartSLOduration=79.899060958 podStartE2EDuration="1m19.899060958s" podCreationTimestamp="2026-02-18 15:22:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:23:19.888718897 +0000 UTC m=+99.274163759" watchObservedRunningTime="2026-02-18 15:23:19.899060958 +0000 UTC m=+99.284505820" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.899307 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-78lnz" podStartSLOduration=78.899303475 podStartE2EDuration="1m18.899303475s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:23:19.899202852 +0000 UTC m=+99.284647714" watchObservedRunningTime="2026-02-18 15:23:19.899303475 +0000 UTC m=+99.284748337" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.910534 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qsjc9" podStartSLOduration=77.910507791 podStartE2EDuration="1m17.910507791s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:23:19.909871473 +0000 UTC m=+99.295316325" watchObservedRunningTime="2026-02-18 15:23:19.910507791 +0000 UTC m=+99.295952653" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.926985 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=72.92696223 podStartE2EDuration="1m12.92696223s" podCreationTimestamp="2026-02-18 15:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:23:19.923784988 +0000 UTC m=+99.309229850" watchObservedRunningTime="2026-02-18 15:23:19.92696223 +0000 UTC m=+99.312407102" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.938966 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" Feb 18 15:23:19 crc kubenswrapper[4968]: I0218 15:23:19.952658 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=50.952641618 podStartE2EDuration="50.952641618s" podCreationTimestamp="2026-02-18 15:22:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:23:19.951438893 +0000 UTC m=+99.336883755" watchObservedRunningTime="2026-02-18 15:23:19.952641618 +0000 UTC m=+99.338086480" Feb 18 15:23:20 crc kubenswrapper[4968]: I0218 15:23:20.229869 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 14:50:50.772720101 +0000 UTC Feb 18 15:23:20 crc kubenswrapper[4968]: I0218 15:23:20.229934 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 18 15:23:20 crc kubenswrapper[4968]: I0218 15:23:20.230090 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:20 crc kubenswrapper[4968]: I0218 15:23:20.230231 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:20 crc kubenswrapper[4968]: E0218 15:23:20.230253 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:20 crc kubenswrapper[4968]: E0218 15:23:20.230394 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:20 crc kubenswrapper[4968]: I0218 15:23:20.238097 4968 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 18 15:23:20 crc kubenswrapper[4968]: I0218 15:23:20.849603 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" event={"ID":"a49d83c1-4f95-4c5b-acd6-e4836ad06be3","Type":"ContainerStarted","Data":"1654057dc2567f93bc4a71db9691e78b282331b31731fd7bb54e618dd8553e22"} Feb 18 15:23:20 crc kubenswrapper[4968]: I0218 15:23:20.849687 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" event={"ID":"a49d83c1-4f95-4c5b-acd6-e4836ad06be3","Type":"ContainerStarted","Data":"f9a661ae67477d381e011a2eb5721c50a8c0fab4ca76681a9adef70a8f60a9f6"} Feb 18 15:23:20 crc kubenswrapper[4968]: I0218 15:23:20.868253 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qq76d" podStartSLOduration=79.868226842 podStartE2EDuration="1m19.868226842s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:23:20.868173041 +0000 UTC m=+100.253617973" watchObservedRunningTime="2026-02-18 15:23:20.868226842 +0000 UTC m=+100.253671704" Feb 18 15:23:20 crc kubenswrapper[4968]: I0218 15:23:20.868851 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podStartSLOduration=79.86884268 podStartE2EDuration="1m19.86884268s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:23:19.966183023 +0000 UTC m=+99.351627885" watchObservedRunningTime="2026-02-18 15:23:20.86884268 +0000 UTC m=+100.254287542" Feb 18 15:23:21 crc kubenswrapper[4968]: I0218 15:23:21.230319 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:21 crc kubenswrapper[4968]: E0218 15:23:21.232608 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:21 crc kubenswrapper[4968]: I0218 15:23:21.232725 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:21 crc kubenswrapper[4968]: E0218 15:23:21.233314 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:21 crc kubenswrapper[4968]: I0218 15:23:21.233539 4968 scope.go:117] "RemoveContainer" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" Feb 18 15:23:21 crc kubenswrapper[4968]: E0218 15:23:21.233707 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" Feb 18 15:23:22 crc kubenswrapper[4968]: I0218 15:23:22.230223 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:22 crc kubenswrapper[4968]: I0218 15:23:22.230313 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:22 crc kubenswrapper[4968]: E0218 15:23:22.230622 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:22 crc kubenswrapper[4968]: E0218 15:23:22.230782 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:23 crc kubenswrapper[4968]: I0218 15:23:23.230657 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:23 crc kubenswrapper[4968]: I0218 15:23:23.230774 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:23 crc kubenswrapper[4968]: E0218 15:23:23.230868 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:23 crc kubenswrapper[4968]: E0218 15:23:23.230959 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:24 crc kubenswrapper[4968]: I0218 15:23:24.230050 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:24 crc kubenswrapper[4968]: I0218 15:23:24.230138 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:24 crc kubenswrapper[4968]: E0218 15:23:24.230904 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:24 crc kubenswrapper[4968]: E0218 15:23:24.230908 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:25 crc kubenswrapper[4968]: I0218 15:23:25.230127 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:25 crc kubenswrapper[4968]: I0218 15:23:25.230213 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:25 crc kubenswrapper[4968]: E0218 15:23:25.230965 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:25 crc kubenswrapper[4968]: E0218 15:23:25.230991 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:26 crc kubenswrapper[4968]: I0218 15:23:26.230046 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:26 crc kubenswrapper[4968]: E0218 15:23:26.230887 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:26 crc kubenswrapper[4968]: I0218 15:23:26.231109 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:26 crc kubenswrapper[4968]: E0218 15:23:26.231190 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:27 crc kubenswrapper[4968]: I0218 15:23:27.230540 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:27 crc kubenswrapper[4968]: I0218 15:23:27.230696 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:27 crc kubenswrapper[4968]: E0218 15:23:27.230966 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:27 crc kubenswrapper[4968]: E0218 15:23:27.231213 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:28 crc kubenswrapper[4968]: I0218 15:23:28.230559 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:28 crc kubenswrapper[4968]: I0218 15:23:28.230726 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:28 crc kubenswrapper[4968]: E0218 15:23:28.230876 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:28 crc kubenswrapper[4968]: E0218 15:23:28.230991 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:29 crc kubenswrapper[4968]: I0218 15:23:29.230438 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:29 crc kubenswrapper[4968]: I0218 15:23:29.230479 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:29 crc kubenswrapper[4968]: E0218 15:23:29.230703 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:29 crc kubenswrapper[4968]: E0218 15:23:29.230831 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:30 crc kubenswrapper[4968]: I0218 15:23:30.230483 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:30 crc kubenswrapper[4968]: I0218 15:23:30.230552 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:30 crc kubenswrapper[4968]: E0218 15:23:30.230693 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:30 crc kubenswrapper[4968]: E0218 15:23:30.230822 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:31 crc kubenswrapper[4968]: I0218 15:23:31.230527 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:31 crc kubenswrapper[4968]: I0218 15:23:31.230667 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:31 crc kubenswrapper[4968]: E0218 15:23:31.232843 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:31 crc kubenswrapper[4968]: E0218 15:23:31.232988 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:32 crc kubenswrapper[4968]: I0218 15:23:32.230502 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:32 crc kubenswrapper[4968]: I0218 15:23:32.230665 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:32 crc kubenswrapper[4968]: E0218 15:23:32.231112 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:32 crc kubenswrapper[4968]: E0218 15:23:32.231250 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:32 crc kubenswrapper[4968]: I0218 15:23:32.231359 4968 scope.go:117] "RemoveContainer" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" Feb 18 15:23:32 crc kubenswrapper[4968]: E0218 15:23:32.231537 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" Feb 18 15:23:33 crc kubenswrapper[4968]: I0218 15:23:33.230016 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:33 crc kubenswrapper[4968]: I0218 15:23:33.230015 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:33 crc kubenswrapper[4968]: E0218 15:23:33.230426 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:33 crc kubenswrapper[4968]: E0218 15:23:33.230362 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:34 crc kubenswrapper[4968]: I0218 15:23:34.230650 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:34 crc kubenswrapper[4968]: I0218 15:23:34.230870 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:34 crc kubenswrapper[4968]: E0218 15:23:34.231079 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:34 crc kubenswrapper[4968]: E0218 15:23:34.231033 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:35 crc kubenswrapper[4968]: I0218 15:23:35.230636 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:35 crc kubenswrapper[4968]: I0218 15:23:35.230725 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:35 crc kubenswrapper[4968]: E0218 15:23:35.230829 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:35 crc kubenswrapper[4968]: E0218 15:23:35.230934 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:36 crc kubenswrapper[4968]: I0218 15:23:36.229878 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:36 crc kubenswrapper[4968]: I0218 15:23:36.229995 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:36 crc kubenswrapper[4968]: E0218 15:23:36.230049 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:36 crc kubenswrapper[4968]: E0218 15:23:36.230209 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:37 crc kubenswrapper[4968]: I0218 15:23:37.229958 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:37 crc kubenswrapper[4968]: I0218 15:23:37.229959 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:37 crc kubenswrapper[4968]: E0218 15:23:37.230154 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:37 crc kubenswrapper[4968]: E0218 15:23:37.230254 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:37 crc kubenswrapper[4968]: I0218 15:23:37.916689 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2qq8_2d01781a-6a78-49a2-80c7-9ac02c810e3f/kube-multus/1.log" Feb 18 15:23:37 crc kubenswrapper[4968]: I0218 15:23:37.917232 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2qq8_2d01781a-6a78-49a2-80c7-9ac02c810e3f/kube-multus/0.log" Feb 18 15:23:37 crc kubenswrapper[4968]: I0218 15:23:37.917290 4968 generic.go:334] "Generic (PLEG): container finished" podID="2d01781a-6a78-49a2-80c7-9ac02c810e3f" containerID="b6ccc51a3ebdf422ba17225bac5935ff711e8ed81619d01a461692c3fb9083eb" exitCode=1 Feb 18 15:23:37 crc kubenswrapper[4968]: I0218 15:23:37.917357 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2qq8" event={"ID":"2d01781a-6a78-49a2-80c7-9ac02c810e3f","Type":"ContainerDied","Data":"b6ccc51a3ebdf422ba17225bac5935ff711e8ed81619d01a461692c3fb9083eb"} Feb 18 15:23:37 crc kubenswrapper[4968]: I0218 15:23:37.917459 4968 scope.go:117] "RemoveContainer" containerID="7f63c69bc2212582aadf808371448f939dea8f977f54861e52c38512b72b81cd" Feb 18 15:23:37 crc kubenswrapper[4968]: I0218 15:23:37.917946 4968 scope.go:117] "RemoveContainer" containerID="b6ccc51a3ebdf422ba17225bac5935ff711e8ed81619d01a461692c3fb9083eb" Feb 18 15:23:37 crc kubenswrapper[4968]: E0218 15:23:37.918120 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-m2qq8_openshift-multus(2d01781a-6a78-49a2-80c7-9ac02c810e3f)\"" pod="openshift-multus/multus-m2qq8" podUID="2d01781a-6a78-49a2-80c7-9ac02c810e3f" Feb 18 15:23:38 crc kubenswrapper[4968]: I0218 15:23:38.229944 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:38 crc kubenswrapper[4968]: I0218 15:23:38.230058 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:38 crc kubenswrapper[4968]: E0218 15:23:38.230136 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:38 crc kubenswrapper[4968]: E0218 15:23:38.230279 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:38 crc kubenswrapper[4968]: I0218 15:23:38.924386 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2qq8_2d01781a-6a78-49a2-80c7-9ac02c810e3f/kube-multus/1.log" Feb 18 15:23:39 crc kubenswrapper[4968]: I0218 15:23:39.229830 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:39 crc kubenswrapper[4968]: I0218 15:23:39.230061 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:39 crc kubenswrapper[4968]: E0218 15:23:39.230264 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:39 crc kubenswrapper[4968]: E0218 15:23:39.230465 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:40 crc kubenswrapper[4968]: I0218 15:23:40.230402 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:40 crc kubenswrapper[4968]: I0218 15:23:40.230402 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:40 crc kubenswrapper[4968]: E0218 15:23:40.230543 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:40 crc kubenswrapper[4968]: E0218 15:23:40.230617 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:41 crc kubenswrapper[4968]: I0218 15:23:41.229908 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:41 crc kubenswrapper[4968]: I0218 15:23:41.229997 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:41 crc kubenswrapper[4968]: E0218 15:23:41.233632 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:41 crc kubenswrapper[4968]: E0218 15:23:41.233796 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:41 crc kubenswrapper[4968]: E0218 15:23:41.252934 4968 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 18 15:23:41 crc kubenswrapper[4968]: E0218 15:23:41.326869 4968 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 18 15:23:42 crc kubenswrapper[4968]: I0218 15:23:42.230242 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:42 crc kubenswrapper[4968]: I0218 15:23:42.230242 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:42 crc kubenswrapper[4968]: E0218 15:23:42.230549 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:42 crc kubenswrapper[4968]: E0218 15:23:42.230645 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:43 crc kubenswrapper[4968]: I0218 15:23:43.230352 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:43 crc kubenswrapper[4968]: E0218 15:23:43.230539 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:43 crc kubenswrapper[4968]: I0218 15:23:43.230658 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:43 crc kubenswrapper[4968]: E0218 15:23:43.230919 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:44 crc kubenswrapper[4968]: I0218 15:23:44.230022 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:44 crc kubenswrapper[4968]: I0218 15:23:44.230085 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:44 crc kubenswrapper[4968]: E0218 15:23:44.230267 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:44 crc kubenswrapper[4968]: E0218 15:23:44.230419 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:45 crc kubenswrapper[4968]: I0218 15:23:45.229647 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:45 crc kubenswrapper[4968]: I0218 15:23:45.229673 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:45 crc kubenswrapper[4968]: E0218 15:23:45.229827 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:45 crc kubenswrapper[4968]: E0218 15:23:45.230105 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:46 crc kubenswrapper[4968]: I0218 15:23:46.229861 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:46 crc kubenswrapper[4968]: I0218 15:23:46.229960 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:46 crc kubenswrapper[4968]: E0218 15:23:46.230357 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:46 crc kubenswrapper[4968]: E0218 15:23:46.230531 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:46 crc kubenswrapper[4968]: I0218 15:23:46.230705 4968 scope.go:117] "RemoveContainer" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" Feb 18 15:23:46 crc kubenswrapper[4968]: E0218 15:23:46.231025 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z2jkk_openshift-ovn-kubernetes(435c6f94-b91e-4ce0-8407-5227f3a5078f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" Feb 18 15:23:46 crc kubenswrapper[4968]: E0218 15:23:46.328098 4968 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 18 15:23:47 crc kubenswrapper[4968]: I0218 15:23:47.230339 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:47 crc kubenswrapper[4968]: I0218 15:23:47.230512 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:47 crc kubenswrapper[4968]: E0218 15:23:47.230608 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:47 crc kubenswrapper[4968]: E0218 15:23:47.230843 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:48 crc kubenswrapper[4968]: I0218 15:23:48.229920 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:48 crc kubenswrapper[4968]: I0218 15:23:48.229994 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:48 crc kubenswrapper[4968]: E0218 15:23:48.230460 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:48 crc kubenswrapper[4968]: E0218 15:23:48.230684 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:49 crc kubenswrapper[4968]: I0218 15:23:49.229954 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:49 crc kubenswrapper[4968]: I0218 15:23:49.230037 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:49 crc kubenswrapper[4968]: E0218 15:23:49.230173 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:49 crc kubenswrapper[4968]: E0218 15:23:49.230373 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:50 crc kubenswrapper[4968]: I0218 15:23:50.230786 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:50 crc kubenswrapper[4968]: I0218 15:23:50.231174 4968 scope.go:117] "RemoveContainer" containerID="b6ccc51a3ebdf422ba17225bac5935ff711e8ed81619d01a461692c3fb9083eb" Feb 18 15:23:50 crc kubenswrapper[4968]: E0218 15:23:50.231234 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:50 crc kubenswrapper[4968]: I0218 15:23:50.232644 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:50 crc kubenswrapper[4968]: E0218 15:23:50.232773 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:50 crc kubenswrapper[4968]: I0218 15:23:50.972617 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2qq8_2d01781a-6a78-49a2-80c7-9ac02c810e3f/kube-multus/1.log" Feb 18 15:23:50 crc kubenswrapper[4968]: I0218 15:23:50.973112 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2qq8" event={"ID":"2d01781a-6a78-49a2-80c7-9ac02c810e3f","Type":"ContainerStarted","Data":"46fd3a8731ea670c4e3883ecc9aa92f701aedd645c1658c7f936d0005d273013"} Feb 18 15:23:51 crc kubenswrapper[4968]: I0218 15:23:51.230675 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:51 crc kubenswrapper[4968]: I0218 15:23:51.231919 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:51 crc kubenswrapper[4968]: E0218 15:23:51.232091 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:51 crc kubenswrapper[4968]: E0218 15:23:51.232945 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:51 crc kubenswrapper[4968]: E0218 15:23:51.329843 4968 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 18 15:23:52 crc kubenswrapper[4968]: I0218 15:23:52.229817 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:52 crc kubenswrapper[4968]: I0218 15:23:52.229906 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:52 crc kubenswrapper[4968]: E0218 15:23:52.229954 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:52 crc kubenswrapper[4968]: E0218 15:23:52.230110 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:53 crc kubenswrapper[4968]: I0218 15:23:53.229797 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:53 crc kubenswrapper[4968]: I0218 15:23:53.229815 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:53 crc kubenswrapper[4968]: E0218 15:23:53.230020 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:53 crc kubenswrapper[4968]: E0218 15:23:53.230158 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:54 crc kubenswrapper[4968]: I0218 15:23:54.229736 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:54 crc kubenswrapper[4968]: I0218 15:23:54.229821 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:54 crc kubenswrapper[4968]: E0218 15:23:54.230059 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:54 crc kubenswrapper[4968]: E0218 15:23:54.230163 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:55 crc kubenswrapper[4968]: I0218 15:23:55.231151 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:55 crc kubenswrapper[4968]: I0218 15:23:55.231256 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:55 crc kubenswrapper[4968]: E0218 15:23:55.231351 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:55 crc kubenswrapper[4968]: E0218 15:23:55.231443 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:56 crc kubenswrapper[4968]: I0218 15:23:56.229632 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:56 crc kubenswrapper[4968]: I0218 15:23:56.229718 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:56 crc kubenswrapper[4968]: E0218 15:23:56.229822 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:56 crc kubenswrapper[4968]: E0218 15:23:56.230018 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:56 crc kubenswrapper[4968]: E0218 15:23:56.331696 4968 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 18 15:23:57 crc kubenswrapper[4968]: I0218 15:23:57.230748 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:57 crc kubenswrapper[4968]: I0218 15:23:57.230875 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:57 crc kubenswrapper[4968]: E0218 15:23:57.230983 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:57 crc kubenswrapper[4968]: E0218 15:23:57.231307 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:58 crc kubenswrapper[4968]: I0218 15:23:58.229809 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:23:58 crc kubenswrapper[4968]: E0218 15:23:58.230014 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:23:58 crc kubenswrapper[4968]: I0218 15:23:58.229822 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:23:58 crc kubenswrapper[4968]: E0218 15:23:58.230241 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:23:59 crc kubenswrapper[4968]: I0218 15:23:59.230284 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:23:59 crc kubenswrapper[4968]: I0218 15:23:59.230368 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:23:59 crc kubenswrapper[4968]: E0218 15:23:59.230808 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:23:59 crc kubenswrapper[4968]: E0218 15:23:59.230980 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:23:59 crc kubenswrapper[4968]: I0218 15:23:59.231245 4968 scope.go:117] "RemoveContainer" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" Feb 18 15:24:00 crc kubenswrapper[4968]: I0218 15:24:00.012050 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/3.log" Feb 18 15:24:00 crc kubenswrapper[4968]: I0218 15:24:00.015435 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerStarted","Data":"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b"} Feb 18 15:24:00 crc kubenswrapper[4968]: I0218 15:24:00.016151 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:24:00 crc kubenswrapper[4968]: I0218 15:24:00.048093 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podStartSLOduration=119.048073681 podStartE2EDuration="1m59.048073681s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:00.047415162 +0000 UTC m=+139.432860044" watchObservedRunningTime="2026-02-18 15:24:00.048073681 +0000 UTC m=+139.433518543" Feb 18 15:24:00 crc kubenswrapper[4968]: I0218 15:24:00.087631 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7sk6k"] Feb 18 15:24:00 crc kubenswrapper[4968]: I0218 15:24:00.087787 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:24:00 crc kubenswrapper[4968]: E0218 15:24:00.087874 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:24:00 crc kubenswrapper[4968]: I0218 15:24:00.230380 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:24:00 crc kubenswrapper[4968]: E0218 15:24:00.230563 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:24:00 crc kubenswrapper[4968]: I0218 15:24:00.230853 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:24:00 crc kubenswrapper[4968]: E0218 15:24:00.230925 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:24:01 crc kubenswrapper[4968]: I0218 15:24:01.229844 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:24:01 crc kubenswrapper[4968]: I0218 15:24:01.229984 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:24:01 crc kubenswrapper[4968]: E0218 15:24:01.231819 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:24:01 crc kubenswrapper[4968]: E0218 15:24:01.232000 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:24:01 crc kubenswrapper[4968]: E0218 15:24:01.333840 4968 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 18 15:24:02 crc kubenswrapper[4968]: I0218 15:24:02.230571 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:24:02 crc kubenswrapper[4968]: I0218 15:24:02.230637 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:24:02 crc kubenswrapper[4968]: E0218 15:24:02.230800 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:24:02 crc kubenswrapper[4968]: E0218 15:24:02.230980 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:24:03 crc kubenswrapper[4968]: I0218 15:24:03.229988 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:24:03 crc kubenswrapper[4968]: I0218 15:24:03.230022 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:24:03 crc kubenswrapper[4968]: E0218 15:24:03.230236 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:24:03 crc kubenswrapper[4968]: E0218 15:24:03.230342 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:24:04 crc kubenswrapper[4968]: I0218 15:24:04.230383 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:24:04 crc kubenswrapper[4968]: E0218 15:24:04.231091 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:24:04 crc kubenswrapper[4968]: I0218 15:24:04.230439 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:24:04 crc kubenswrapper[4968]: E0218 15:24:04.231334 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:24:05 crc kubenswrapper[4968]: I0218 15:24:05.230144 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:24:05 crc kubenswrapper[4968]: I0218 15:24:05.230184 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:24:05 crc kubenswrapper[4968]: E0218 15:24:05.230550 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 15:24:05 crc kubenswrapper[4968]: E0218 15:24:05.230678 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7sk6k" podUID="9185b44b-bac7-458b-b4d9-4c389da07c14" Feb 18 15:24:06 crc kubenswrapper[4968]: I0218 15:24:06.230523 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:24:06 crc kubenswrapper[4968]: I0218 15:24:06.230660 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:24:06 crc kubenswrapper[4968]: E0218 15:24:06.230699 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 15:24:06 crc kubenswrapper[4968]: E0218 15:24:06.230920 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 15:24:07 crc kubenswrapper[4968]: I0218 15:24:07.231640 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:24:07 crc kubenswrapper[4968]: I0218 15:24:07.231795 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:24:07 crc kubenswrapper[4968]: I0218 15:24:07.236650 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 18 15:24:07 crc kubenswrapper[4968]: I0218 15:24:07.236717 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 18 15:24:07 crc kubenswrapper[4968]: I0218 15:24:07.236650 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 18 15:24:07 crc kubenswrapper[4968]: I0218 15:24:07.238314 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 18 15:24:08 crc kubenswrapper[4968]: I0218 15:24:08.230611 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:24:08 crc kubenswrapper[4968]: I0218 15:24:08.230696 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:24:08 crc kubenswrapper[4968]: I0218 15:24:08.233653 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 18 15:24:08 crc kubenswrapper[4968]: I0218 15:24:08.234683 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 18 15:24:09 crc kubenswrapper[4968]: I0218 15:24:09.246670 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:09 crc kubenswrapper[4968]: E0218 15:24:09.247008 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:26:11.246965023 +0000 UTC m=+270.632409885 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:09 crc kubenswrapper[4968]: I0218 15:24:09.347865 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:24:09 crc kubenswrapper[4968]: I0218 15:24:09.347938 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:24:09 crc kubenswrapper[4968]: I0218 15:24:09.347966 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:24:09 crc kubenswrapper[4968]: I0218 15:24:09.347998 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:24:09 crc kubenswrapper[4968]: I0218 15:24:09.349309 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:24:09 crc kubenswrapper[4968]: I0218 15:24:09.355661 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:24:09 crc kubenswrapper[4968]: I0218 15:24:09.356139 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:24:09 crc kubenswrapper[4968]: I0218 15:24:09.356662 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:24:09 crc kubenswrapper[4968]: I0218 15:24:09.374805 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 15:24:09 crc kubenswrapper[4968]: I0218 15:24:09.454070 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:24:09 crc kubenswrapper[4968]: I0218 15:24:09.463699 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 15:24:09 crc kubenswrapper[4968]: W0218 15:24:09.685669 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-b0ef955d759e062bd7c01f50fa331ea1798c4ef86d52cd68e65f4e0ac2189f55 WatchSource:0}: Error finding container b0ef955d759e062bd7c01f50fa331ea1798c4ef86d52cd68e65f4e0ac2189f55: Status 404 returned error can't find the container with id b0ef955d759e062bd7c01f50fa331ea1798c4ef86d52cd68e65f4e0ac2189f55 Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.057351 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"dc6afbe7dae71bce0614c7a0ac12b99caba768eb8ce54a4aa3f4319812f965da"} Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.057482 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f2b8135189609a35034672ae83dc05be77205f928688e3d12102ad0debc60388"} Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.058446 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.058668 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6720682883a5739f001da78c19dcaffe9314ece3c838042da96fe42ebb168f09"} Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.058771 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b0ef955d759e062bd7c01f50fa331ea1798c4ef86d52cd68e65f4e0ac2189f55"} Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.060093 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6e638cd18d46df568bc04c67ad8fd2819ec1bd94abc64a6b958e5e2e79c82c88"} Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.060325 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"f45dd6efa668e2a9a508333d7a9a942f0c4d844a018b80d687abadc49ea692dd"} Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.682923 4968 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.722782 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.723549 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.730850 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.730883 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.730889 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.731496 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.732379 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.732549 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.732629 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.732637 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.733098 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.735440 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.736080 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.748702 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-m6p2k"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.749361 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-m6p2k" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.750785 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.750999 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.751621 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.767208 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.782227 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.784633 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6v575"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.784802 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/03d43fa6-7663-45ca-b028-0d245cc505af-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mjtgm\" (UID: \"03d43fa6-7663-45ca-b028-0d245cc505af\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.784858 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ecce765-efd2-479b-9512-295892b014a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.784893 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5ecce765-efd2-479b-9512-295892b014a7-etcd-client\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.784968 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kfgk\" (UniqueName: \"kubernetes.io/projected/a8b4bf38-6d44-4b8a-8c08-afa0023e6d28-kube-api-access-4kfgk\") pod \"dns-operator-744455d44c-m6p2k\" (UID: \"a8b4bf38-6d44-4b8a-8c08-afa0023e6d28\") " pod="openshift-dns-operator/dns-operator-744455d44c-m6p2k" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785042 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5ecce765-efd2-479b-9512-295892b014a7-audit-dir\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785094 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5ecce765-efd2-479b-9512-295892b014a7-audit-policies\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785128 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw4nc\" (UniqueName: \"kubernetes.io/projected/03d43fa6-7663-45ca-b028-0d245cc505af-kube-api-access-kw4nc\") pod \"cluster-image-registry-operator-dc59b4c8b-mjtgm\" (UID: \"03d43fa6-7663-45ca-b028-0d245cc505af\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785159 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a8b4bf38-6d44-4b8a-8c08-afa0023e6d28-metrics-tls\") pod \"dns-operator-744455d44c-m6p2k\" (UID: \"a8b4bf38-6d44-4b8a-8c08-afa0023e6d28\") " pod="openshift-dns-operator/dns-operator-744455d44c-m6p2k" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785258 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785282 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5ecce765-efd2-479b-9512-295892b014a7-encryption-config\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785332 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/03d43fa6-7663-45ca-b028-0d245cc505af-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mjtgm\" (UID: \"03d43fa6-7663-45ca-b028-0d245cc505af\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785357 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/925dd643-0943-43dd-a71e-c7cbcfae2c00-serving-cert\") pod \"openshift-config-operator-7777fb866f-ckl6x\" (UID: \"925dd643-0943-43dd-a71e-c7cbcfae2c00\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785382 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/03d43fa6-7663-45ca-b028-0d245cc505af-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mjtgm\" (UID: \"03d43fa6-7663-45ca-b028-0d245cc505af\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785509 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqtpd\" (UniqueName: \"kubernetes.io/projected/925dd643-0943-43dd-a71e-c7cbcfae2c00-kube-api-access-zqtpd\") pod \"openshift-config-operator-7777fb866f-ckl6x\" (UID: \"925dd643-0943-43dd-a71e-c7cbcfae2c00\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785594 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4bk8\" (UniqueName: \"kubernetes.io/projected/5ecce765-efd2-479b-9512-295892b014a7-kube-api-access-v4bk8\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785631 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/925dd643-0943-43dd-a71e-c7cbcfae2c00-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ckl6x\" (UID: \"925dd643-0943-43dd-a71e-c7cbcfae2c00\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785661 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5ecce765-efd2-479b-9512-295892b014a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785688 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ecce765-efd2-479b-9512-295892b014a7-serving-cert\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785736 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-h58tr"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.785744 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.786025 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.786183 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.786364 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.787335 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.788247 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7b762"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.788846 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.789497 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.789613 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.790537 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.794212 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.794709 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.795360 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.797087 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-s6qsc"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.797872 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.799133 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.799201 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.799339 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.799408 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.799641 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.799695 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.800477 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.806564 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.807507 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.808241 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-whmt5"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.809234 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-td7d6"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.809697 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-td7d6" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.810043 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.810831 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.810927 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.810831 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.811038 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.810867 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.810890 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.811240 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.811507 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.811555 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.811607 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.812085 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.812241 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.812372 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.812467 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.812591 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.812489 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.812705 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.812780 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.812903 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.812953 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.813082 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.814567 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.814647 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.814782 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.814906 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.815292 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.816609 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nkbhx"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.817120 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.833341 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.834203 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.834673 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nz8lf"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.835273 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-nz8lf" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.839383 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.839707 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.839732 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.839904 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.839921 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.839999 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840072 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840184 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840301 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840306 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.839581 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840023 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840534 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840641 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840651 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840717 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840775 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840783 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840852 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.840856 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.839537 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.841063 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.841645 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.888606 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.889295 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.891202 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.893092 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.893532 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.896035 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.897654 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.897830 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41a0f2d1-405f-4396-a081-b874a35bacf8-serving-cert\") pod \"console-operator-58897d9998-nkbhx\" (UID: \"41a0f2d1-405f-4396-a081-b874a35bacf8\") " pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.897870 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/402543c2-2831-4d3b-87ca-6d284c6d9be1-images\") pod \"machine-api-operator-5694c8668f-h58tr\" (UID: \"402543c2-2831-4d3b-87ca-6d284c6d9be1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.897910 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv8hk\" (UniqueName: \"kubernetes.io/projected/402543c2-2831-4d3b-87ca-6d284c6d9be1-kube-api-access-xv8hk\") pod \"machine-api-operator-5694c8668f-h58tr\" (UID: \"402543c2-2831-4d3b-87ca-6d284c6d9be1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.897931 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41a0f2d1-405f-4396-a081-b874a35bacf8-config\") pod \"console-operator-58897d9998-nkbhx\" (UID: \"41a0f2d1-405f-4396-a081-b874a35bacf8\") " pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.897954 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-etcd-service-ca\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.897983 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kfgk\" (UniqueName: \"kubernetes.io/projected/a8b4bf38-6d44-4b8a-8c08-afa0023e6d28-kube-api-access-4kfgk\") pod \"dns-operator-744455d44c-m6p2k\" (UID: \"a8b4bf38-6d44-4b8a-8c08-afa0023e6d28\") " pod="openshift-dns-operator/dns-operator-744455d44c-m6p2k" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898016 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5ecce765-efd2-479b-9512-295892b014a7-audit-dir\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898038 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5ecce765-efd2-479b-9512-295892b014a7-audit-policies\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898058 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d3bb5-c13d-49fc-9d5c-5b2ce0847d92-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4q2wd\" (UID: \"210d3bb5-c13d-49fc-9d5c-5b2ce0847d92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898081 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw4nc\" (UniqueName: \"kubernetes.io/projected/03d43fa6-7663-45ca-b028-0d245cc505af-kube-api-access-kw4nc\") pod \"cluster-image-registry-operator-dc59b4c8b-mjtgm\" (UID: \"03d43fa6-7663-45ca-b028-0d245cc505af\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898101 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a8b4bf38-6d44-4b8a-8c08-afa0023e6d28-metrics-tls\") pod \"dns-operator-744455d44c-m6p2k\" (UID: \"a8b4bf38-6d44-4b8a-8c08-afa0023e6d28\") " pod="openshift-dns-operator/dns-operator-744455d44c-m6p2k" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898121 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-config\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898143 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d43c1374-e913-423e-ab78-6e14b1c16bfd-config\") pod \"machine-approver-56656f9798-hjsr5\" (UID: \"d43c1374-e913-423e-ab78-6e14b1c16bfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898160 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f57a4a48-39f0-44f3-bbf8-982a27fd6f14-apiservice-cert\") pod \"packageserver-d55dfcdfc-7kwbt\" (UID: \"f57a4a48-39f0-44f3-bbf8-982a27fd6f14\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898183 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5ecce765-efd2-479b-9512-295892b014a7-encryption-config\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898205 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz4k2\" (UniqueName: \"kubernetes.io/projected/41a0f2d1-405f-4396-a081-b874a35bacf8-kube-api-access-dz4k2\") pod \"console-operator-58897d9998-nkbhx\" (UID: \"41a0f2d1-405f-4396-a081-b874a35bacf8\") " pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898224 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-client-ca\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898264 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vljcg\" (UniqueName: \"kubernetes.io/projected/d8e923a2-9b5c-49db-8a02-5eff8ca31033-kube-api-access-vljcg\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898320 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/03d43fa6-7663-45ca-b028-0d245cc505af-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mjtgm\" (UID: \"03d43fa6-7663-45ca-b028-0d245cc505af\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898346 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/03d43fa6-7663-45ca-b028-0d245cc505af-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mjtgm\" (UID: \"03d43fa6-7663-45ca-b028-0d245cc505af\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898369 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/925dd643-0943-43dd-a71e-c7cbcfae2c00-serving-cert\") pod \"openshift-config-operator-7777fb866f-ckl6x\" (UID: \"925dd643-0943-43dd-a71e-c7cbcfae2c00\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898388 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e923a2-9b5c-49db-8a02-5eff8ca31033-serving-cert\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898410 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvf2k\" (UniqueName: \"kubernetes.io/projected/d43c1374-e913-423e-ab78-6e14b1c16bfd-kube-api-access-nvf2k\") pod \"machine-approver-56656f9798-hjsr5\" (UID: \"d43c1374-e913-423e-ab78-6e14b1c16bfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898429 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-etcd-ca\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898454 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqtpd\" (UniqueName: \"kubernetes.io/projected/925dd643-0943-43dd-a71e-c7cbcfae2c00-kube-api-access-zqtpd\") pod \"openshift-config-operator-7777fb866f-ckl6x\" (UID: \"925dd643-0943-43dd-a71e-c7cbcfae2c00\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898471 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/402543c2-2831-4d3b-87ca-6d284c6d9be1-config\") pod \"machine-api-operator-5694c8668f-h58tr\" (UID: \"402543c2-2831-4d3b-87ca-6d284c6d9be1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898491 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d3bb5-c13d-49fc-9d5c-5b2ce0847d92-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4q2wd\" (UID: \"210d3bb5-c13d-49fc-9d5c-5b2ce0847d92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898514 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5ecce765-efd2-479b-9512-295892b014a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898537 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4bk8\" (UniqueName: \"kubernetes.io/projected/5ecce765-efd2-479b-9512-295892b014a7-kube-api-access-v4bk8\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898557 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/925dd643-0943-43dd-a71e-c7cbcfae2c00-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ckl6x\" (UID: \"925dd643-0943-43dd-a71e-c7cbcfae2c00\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898579 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qh7t\" (UniqueName: \"kubernetes.io/projected/210d3bb5-c13d-49fc-9d5c-5b2ce0847d92-kube-api-access-5qh7t\") pod \"openshift-controller-manager-operator-756b6f6bc6-4q2wd\" (UID: \"210d3bb5-c13d-49fc-9d5c-5b2ce0847d92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898597 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-serving-cert\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898618 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ecce765-efd2-479b-9512-295892b014a7-serving-cert\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898658 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/03d43fa6-7663-45ca-b028-0d245cc505af-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mjtgm\" (UID: \"03d43fa6-7663-45ca-b028-0d245cc505af\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.898678 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.907036 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5ecce765-efd2-479b-9512-295892b014a7-audit-dir\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.907840 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5ecce765-efd2-479b-9512-295892b014a7-audit-policies\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.909685 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/402543c2-2831-4d3b-87ca-6d284c6d9be1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-h58tr\" (UID: \"402543c2-2831-4d3b-87ca-6d284c6d9be1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.909758 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmrpc\" (UniqueName: \"kubernetes.io/projected/f57a4a48-39f0-44f3-bbf8-982a27fd6f14-kube-api-access-bmrpc\") pod \"packageserver-d55dfcdfc-7kwbt\" (UID: \"f57a4a48-39f0-44f3-bbf8-982a27fd6f14\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.911243 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.909801 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d43c1374-e913-423e-ab78-6e14b1c16bfd-machine-approver-tls\") pod \"machine-approver-56656f9798-hjsr5\" (UID: \"d43c1374-e913-423e-ab78-6e14b1c16bfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.951077 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f57a4a48-39f0-44f3-bbf8-982a27fd6f14-webhook-cert\") pod \"packageserver-d55dfcdfc-7kwbt\" (UID: \"f57a4a48-39f0-44f3-bbf8-982a27fd6f14\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.951159 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ecce765-efd2-479b-9512-295892b014a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.951191 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-config\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.951213 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f57a4a48-39f0-44f3-bbf8-982a27fd6f14-tmpfs\") pod \"packageserver-d55dfcdfc-7kwbt\" (UID: \"f57a4a48-39f0-44f3-bbf8-982a27fd6f14\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.951247 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5ecce765-efd2-479b-9512-295892b014a7-etcd-client\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.951284 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d43c1374-e913-423e-ab78-6e14b1c16bfd-auth-proxy-config\") pod \"machine-approver-56656f9798-hjsr5\" (UID: \"d43c1374-e913-423e-ab78-6e14b1c16bfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.951321 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41a0f2d1-405f-4396-a081-b874a35bacf8-trusted-ca\") pod \"console-operator-58897d9998-nkbhx\" (UID: \"41a0f2d1-405f-4396-a081-b874a35bacf8\") " pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.951351 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-etcd-client\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.951371 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/925dd643-0943-43dd-a71e-c7cbcfae2c00-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ckl6x\" (UID: \"925dd643-0943-43dd-a71e-c7cbcfae2c00\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.953160 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ecce765-efd2-479b-9512-295892b014a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.953382 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.953618 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5ecce765-efd2-479b-9512-295892b014a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.953724 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.954295 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.962330 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ecce765-efd2-479b-9512-295892b014a7-serving-cert\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.962329 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/03d43fa6-7663-45ca-b028-0d245cc505af-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mjtgm\" (UID: \"03d43fa6-7663-45ca-b028-0d245cc505af\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.963794 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.963988 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.964115 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.964305 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.964465 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.964614 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.965654 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.967255 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.967373 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/925dd643-0943-43dd-a71e-c7cbcfae2c00-serving-cert\") pod \"openshift-config-operator-7777fb866f-ckl6x\" (UID: \"925dd643-0943-43dd-a71e-c7cbcfae2c00\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.967396 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.967518 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.967637 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5ecce765-efd2-479b-9512-295892b014a7-encryption-config\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.968692 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.968890 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a8b4bf38-6d44-4b8a-8c08-afa0023e6d28-metrics-tls\") pod \"dns-operator-744455d44c-m6p2k\" (UID: \"a8b4bf38-6d44-4b8a-8c08-afa0023e6d28\") " pod="openshift-dns-operator/dns-operator-744455d44c-m6p2k" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.953896 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.968972 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/03d43fa6-7663-45ca-b028-0d245cc505af-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mjtgm\" (UID: \"03d43fa6-7663-45ca-b028-0d245cc505af\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.969000 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.969136 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.969148 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.969175 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.969269 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.969365 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.969654 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.969925 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.969139 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.970893 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.971293 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.971581 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.971727 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.969277 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.976469 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw4nc\" (UniqueName: \"kubernetes.io/projected/03d43fa6-7663-45ca-b028-0d245cc505af-kube-api-access-kw4nc\") pod \"cluster-image-registry-operator-dc59b4c8b-mjtgm\" (UID: \"03d43fa6-7663-45ca-b028-0d245cc505af\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.979000 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5ecce765-efd2-479b-9512-295892b014a7-etcd-client\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.985599 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqtpd\" (UniqueName: \"kubernetes.io/projected/925dd643-0943-43dd-a71e-c7cbcfae2c00-kube-api-access-zqtpd\") pod \"openshift-config-operator-7777fb866f-ckl6x\" (UID: \"925dd643-0943-43dd-a71e-c7cbcfae2c00\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.985981 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.986105 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.985979 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.986585 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dnmb8"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.987077 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-857s6"] Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.987485 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.987876 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.988078 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.989761 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4bk8\" (UniqueName: \"kubernetes.io/projected/5ecce765-efd2-479b-9512-295892b014a7-kube-api-access-v4bk8\") pod \"apiserver-7bbb656c7d-qplwp\" (UID: \"5ecce765-efd2-479b-9512-295892b014a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.990074 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.990125 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.990383 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.990519 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.990824 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.990957 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.993332 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 18 15:24:10 crc kubenswrapper[4968]: I0218 15:24:10.999181 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kfgk\" (UniqueName: \"kubernetes.io/projected/a8b4bf38-6d44-4b8a-8c08-afa0023e6d28-kube-api-access-4kfgk\") pod \"dns-operator-744455d44c-m6p2k\" (UID: \"a8b4bf38-6d44-4b8a-8c08-afa0023e6d28\") " pod="openshift-dns-operator/dns-operator-744455d44c-m6p2k" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.003668 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.004357 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.007280 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/03d43fa6-7663-45ca-b028-0d245cc505af-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mjtgm\" (UID: \"03d43fa6-7663-45ca-b028-0d245cc505af\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.007394 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dxl4k"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.011566 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8g9pk"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.012035 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lcx97"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.012431 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.012718 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.012903 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.016704 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-x829v"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.017826 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.017933 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.022564 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.023786 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.029449 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.034087 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-w9zh9"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.035375 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.036469 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.037716 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.039041 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.044672 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.045802 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.046523 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6v575"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.048431 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.048960 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-796c7"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.065959 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.071697 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-lmz7d"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.072486 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-796c7" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.073046 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-h58tr"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.073071 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.073090 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.073627 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.075183 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-encryption-config\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.075266 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-service-ca\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.075329 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-image-import-ca\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.075395 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfffde2b-573c-418b-903b-9117e958dabc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mncr8\" (UID: \"dfffde2b-573c-418b-903b-9117e958dabc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.075648 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39580550-6421-4eb7-8727-42c26ad5c3f2-serving-cert\") pod \"route-controller-manager-6576b87f9c-f7s66\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.075699 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtghp\" (UniqueName: \"kubernetes.io/projected/954a7c44-9845-4a84-8cd6-a5fe9be2d7d2-kube-api-access-jtghp\") pod \"machine-config-controller-84d6567774-x829v\" (UID: \"954a7c44-9845-4a84-8cd6-a5fe9be2d7d2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.075732 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d43c1374-e913-423e-ab78-6e14b1c16bfd-auth-proxy-config\") pod \"machine-approver-56656f9798-hjsr5\" (UID: \"d43c1374-e913-423e-ab78-6e14b1c16bfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.075826 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-config\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.076124 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th4hp\" (UniqueName: \"kubernetes.io/projected/ba1be0b0-ffa7-4b12-bc99-354621c8dd5e-kube-api-access-th4hp\") pod \"multus-admission-controller-857f4d67dd-nz8lf\" (UID: \"ba1be0b0-ffa7-4b12-bc99-354621c8dd5e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nz8lf" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.076628 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dggsh\" (UniqueName: \"kubernetes.io/projected/6c8c2a1f-4652-4791-a114-ebb9a46659ec-kube-api-access-dggsh\") pod \"cluster-samples-operator-665b6dd947-q5lx6\" (UID: \"6c8c2a1f-4652-4791-a114-ebb9a46659ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.076880 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1c808375-740d-4fe5-ae91-64f776b28e8a-metrics-certs\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.076989 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-oauth-serving-cert\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.077046 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dnmb8\" (UID: \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.077113 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.077214 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.077238 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d43c1374-e913-423e-ab78-6e14b1c16bfd-auth-proxy-config\") pod \"machine-approver-56656f9798-hjsr5\" (UID: \"d43c1374-e913-423e-ab78-6e14b1c16bfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.077348 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.077447 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41a0f2d1-405f-4396-a081-b874a35bacf8-trusted-ca\") pod \"console-operator-58897d9998-nkbhx\" (UID: \"41a0f2d1-405f-4396-a081-b874a35bacf8\") " pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.078078 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2c7da228-335d-4f50-83e6-4ae152ed4874-console-oauth-config\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.080172 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-audit-dir\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.080311 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.080412 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-etcd-client\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.080482 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-etcd-client\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.080548 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-console-config\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.081931 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41a0f2d1-405f-4396-a081-b874a35bacf8-trusted-ca\") pod \"console-operator-58897d9998-nkbhx\" (UID: \"41a0f2d1-405f-4396-a081-b874a35bacf8\") " pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.082128 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-h695r"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.082317 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.082481 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41a0f2d1-405f-4396-a081-b874a35bacf8-serving-cert\") pod \"console-operator-58897d9998-nkbhx\" (UID: \"41a0f2d1-405f-4396-a081-b874a35bacf8\") " pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.082618 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/402543c2-2831-4d3b-87ca-6d284c6d9be1-images\") pod \"machine-api-operator-5694c8668f-h58tr\" (UID: \"402543c2-2831-4d3b-87ca-6d284c6d9be1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.082697 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv8hk\" (UniqueName: \"kubernetes.io/projected/402543c2-2831-4d3b-87ca-6d284c6d9be1-kube-api-access-xv8hk\") pod \"machine-api-operator-5694c8668f-h58tr\" (UID: \"402543c2-2831-4d3b-87ca-6d284c6d9be1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.082815 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39580550-6421-4eb7-8727-42c26ad5c3f2-config\") pod \"route-controller-manager-6576b87f9c-f7s66\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.082972 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.083106 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41a0f2d1-405f-4396-a081-b874a35bacf8-config\") pod \"console-operator-58897d9998-nkbhx\" (UID: \"41a0f2d1-405f-4396-a081-b874a35bacf8\") " pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.083182 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-etcd-service-ca\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.083263 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85chc\" (UniqueName: \"kubernetes.io/projected/39580550-6421-4eb7-8727-42c26ad5c3f2-kube-api-access-85chc\") pod \"route-controller-manager-6576b87f9c-f7s66\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.083341 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjnc7\" (UniqueName: \"kubernetes.io/projected/605d503b-946e-47e6-b08f-c6eaf0c3eb9a-kube-api-access-pjnc7\") pod \"control-plane-machine-set-operator-78cbb6b69f-x9mtt\" (UID: \"605d503b-946e-47e6-b08f-c6eaf0c3eb9a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.083437 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/29f23275-ee4d-41d6-acaa-a9c1363d71c3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p58wk\" (UID: \"29f23275-ee4d-41d6-acaa-a9c1363d71c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.083514 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dnmb8\" (UID: \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.083586 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-serving-cert\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.083656 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ba1be0b0-ffa7-4b12-bc99-354621c8dd5e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nz8lf\" (UID: \"ba1be0b0-ffa7-4b12-bc99-354621c8dd5e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nz8lf" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.083727 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bdb30229-c0c5-4f40-a69c-6d8548cb76d8-profile-collector-cert\") pod \"catalog-operator-68c6474976-cx7j6\" (UID: \"bdb30229-c0c5-4f40-a69c-6d8548cb76d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.083816 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1c808375-740d-4fe5-ae91-64f776b28e8a-stats-auth\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.083886 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-audit-policies\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.083958 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57795f27-c9c3-4f38-a5da-364e2488f343-audit-dir\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.084899 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d3bb5-c13d-49fc-9d5c-5b2ce0847d92-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4q2wd\" (UID: \"210d3bb5-c13d-49fc-9d5c-5b2ce0847d92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.085021 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-etcd-serving-ca\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.085105 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.085184 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d43c1374-e913-423e-ab78-6e14b1c16bfd-config\") pod \"machine-approver-56656f9798-hjsr5\" (UID: \"d43c1374-e913-423e-ab78-6e14b1c16bfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.085291 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-config\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.085373 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w7n5\" (UniqueName: \"kubernetes.io/projected/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-kube-api-access-9w7n5\") pod \"marketplace-operator-79b997595-dnmb8\" (UID: \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.085451 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f57a4a48-39f0-44f3-bbf8-982a27fd6f14-apiservice-cert\") pod \"packageserver-d55dfcdfc-7kwbt\" (UID: \"f57a4a48-39f0-44f3-bbf8-982a27fd6f14\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.085551 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-node-pullsecrets\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.085625 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfffde2b-573c-418b-903b-9117e958dabc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mncr8\" (UID: \"dfffde2b-573c-418b-903b-9117e958dabc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.085692 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df2ba163-1935-4600-a6c8-dd73d151ed32-secret-volume\") pod \"collect-profiles-29523795-hd6q4\" (UID: \"df2ba163-1935-4600-a6c8-dd73d151ed32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.085787 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz4k2\" (UniqueName: \"kubernetes.io/projected/41a0f2d1-405f-4396-a081-b874a35bacf8-kube-api-access-dz4k2\") pod \"console-operator-58897d9998-nkbhx\" (UID: \"41a0f2d1-405f-4396-a081-b874a35bacf8\") " pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.085868 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c7da228-335d-4f50-83e6-4ae152ed4874-console-serving-cert\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.085947 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxx4s\" (UniqueName: \"kubernetes.io/projected/2c7da228-335d-4f50-83e6-4ae152ed4874-kube-api-access-dxx4s\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.086013 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/29f23275-ee4d-41d6-acaa-a9c1363d71c3-srv-cert\") pod \"olm-operator-6b444d44fb-p58wk\" (UID: \"29f23275-ee4d-41d6-acaa-a9c1363d71c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.086088 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.086165 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-client-ca\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.086232 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vljcg\" (UniqueName: \"kubernetes.io/projected/d8e923a2-9b5c-49db-8a02-5eff8ca31033-kube-api-access-vljcg\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.086304 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76sj5\" (UniqueName: \"kubernetes.io/projected/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-kube-api-access-76sj5\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.086368 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-audit\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.087469 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/402543c2-2831-4d3b-87ca-6d284c6d9be1-images\") pod \"machine-api-operator-5694c8668f-h58tr\" (UID: \"402543c2-2831-4d3b-87ca-6d284c6d9be1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.088094 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41a0f2d1-405f-4396-a081-b874a35bacf8-config\") pod \"console-operator-58897d9998-nkbhx\" (UID: \"41a0f2d1-405f-4396-a081-b874a35bacf8\") " pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.089086 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d3bb5-c13d-49fc-9d5c-5b2ce0847d92-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4q2wd\" (UID: \"210d3bb5-c13d-49fc-9d5c-5b2ce0847d92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.089511 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d43c1374-e913-423e-ab78-6e14b1c16bfd-config\") pod \"machine-approver-56656f9798-hjsr5\" (UID: \"d43c1374-e913-423e-ab78-6e14b1c16bfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.090067 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-config\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.086443 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrwtn\" (UniqueName: \"kubernetes.io/projected/df2ba163-1935-4600-a6c8-dd73d151ed32-kube-api-access-lrwtn\") pod \"collect-profiles-29523795-hd6q4\" (UID: \"df2ba163-1935-4600-a6c8-dd73d151ed32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.091219 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.091294 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39580550-6421-4eb7-8727-42c26ad5c3f2-client-ca\") pod \"route-controller-manager-6576b87f9c-f7s66\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.091362 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zrzs\" (UniqueName: \"kubernetes.io/projected/bdb30229-c0c5-4f40-a69c-6d8548cb76d8-kube-api-access-2zrzs\") pod \"catalog-operator-68c6474976-cx7j6\" (UID: \"bdb30229-c0c5-4f40-a69c-6d8548cb76d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.091437 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.092063 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.092326 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-etcd-service-ca\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.092984 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/605d503b-946e-47e6-b08f-c6eaf0c3eb9a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-x9mtt\" (UID: \"605d503b-946e-47e6-b08f-c6eaf0c3eb9a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.093395 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.093682 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-h695r" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.093796 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-client-ca\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.093854 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-etcd-ca\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.100303 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.094283 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-etcd-ca\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.096133 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41a0f2d1-405f-4396-a081-b874a35bacf8-serving-cert\") pod \"console-operator-58897d9998-nkbhx\" (UID: \"41a0f2d1-405f-4396-a081-b874a35bacf8\") " pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.096194 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f57a4a48-39f0-44f3-bbf8-982a27fd6f14-apiservice-cert\") pod \"packageserver-d55dfcdfc-7kwbt\" (UID: \"f57a4a48-39f0-44f3-bbf8-982a27fd6f14\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.096221 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-m6p2k"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.100474 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.100491 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.098121 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.095621 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-etcd-client\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.100820 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e923a2-9b5c-49db-8a02-5eff8ca31033-serving-cert\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.100983 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvf2k\" (UniqueName: \"kubernetes.io/projected/d43c1374-e913-423e-ab78-6e14b1c16bfd-kube-api-access-nvf2k\") pod \"machine-approver-56656f9798-hjsr5\" (UID: \"d43c1374-e913-423e-ab78-6e14b1c16bfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.101070 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djw7b\" (UniqueName: \"kubernetes.io/projected/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-kube-api-access-djw7b\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.101156 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/954a7c44-9845-4a84-8cd6-a5fe9be2d7d2-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-x829v\" (UID: \"954a7c44-9845-4a84-8cd6-a5fe9be2d7d2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.101314 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c808375-740d-4fe5-ae91-64f776b28e8a-service-ca-bundle\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.101397 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df2ba163-1935-4600-a6c8-dd73d151ed32-config-volume\") pod \"collect-profiles-29523795-hd6q4\" (UID: \"df2ba163-1935-4600-a6c8-dd73d151ed32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.101472 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.101554 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrwvb\" (UniqueName: \"kubernetes.io/projected/29f23275-ee4d-41d6-acaa-a9c1363d71c3-kube-api-access-lrwvb\") pod \"olm-operator-6b444d44fb-p58wk\" (UID: \"29f23275-ee4d-41d6-acaa-a9c1363d71c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.101634 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/402543c2-2831-4d3b-87ca-6d284c6d9be1-config\") pod \"machine-api-operator-5694c8668f-h58tr\" (UID: \"402543c2-2831-4d3b-87ca-6d284c6d9be1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.101704 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d3bb5-c13d-49fc-9d5c-5b2ce0847d92-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4q2wd\" (UID: \"210d3bb5-c13d-49fc-9d5c-5b2ce0847d92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.101795 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-trusted-ca-bundle\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.101869 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/954a7c44-9845-4a84-8cd6-a5fe9be2d7d2-proxy-tls\") pod \"machine-config-controller-84d6567774-x829v\" (UID: \"954a7c44-9845-4a84-8cd6-a5fe9be2d7d2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.101956 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntt9n\" (UniqueName: \"kubernetes.io/projected/1c808375-740d-4fe5-ae91-64f776b28e8a-kube-api-access-ntt9n\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.102045 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bdb30229-c0c5-4f40-a69c-6d8548cb76d8-srv-cert\") pod \"catalog-operator-68c6474976-cx7j6\" (UID: \"bdb30229-c0c5-4f40-a69c-6d8548cb76d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.102141 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qh7t\" (UniqueName: \"kubernetes.io/projected/210d3bb5-c13d-49fc-9d5c-5b2ce0847d92-kube-api-access-5qh7t\") pod \"openshift-controller-manager-operator-756b6f6bc6-4q2wd\" (UID: \"210d3bb5-c13d-49fc-9d5c-5b2ce0847d92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.102501 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/402543c2-2831-4d3b-87ca-6d284c6d9be1-config\") pod \"machine-api-operator-5694c8668f-h58tr\" (UID: \"402543c2-2831-4d3b-87ca-6d284c6d9be1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.102628 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-serving-cert\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.102913 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-nhsgq\" (UID: \"4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.102995 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1c808375-740d-4fe5-ae91-64f776b28e8a-default-certificate\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103187 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103277 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-969jw\" (UniqueName: \"kubernetes.io/projected/57795f27-c9c3-4f38-a5da-364e2488f343-kube-api-access-969jw\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103352 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmmjt\" (UniqueName: \"kubernetes.io/projected/5985084a-4c95-41f7-a40a-50550738e31c-kube-api-access-fmmjt\") pod \"downloads-7954f5f757-td7d6\" (UID: \"5985084a-4c95-41f7-a40a-50550738e31c\") " pod="openshift-console/downloads-7954f5f757-td7d6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103507 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-trusted-ca-bundle\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103035 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nkbhx"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103775 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-td7d6"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103676 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103865 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/402543c2-2831-4d3b-87ca-6d284c6d9be1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-h58tr\" (UID: \"402543c2-2831-4d3b-87ca-6d284c6d9be1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103887 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmrpc\" (UniqueName: \"kubernetes.io/projected/f57a4a48-39f0-44f3-bbf8-982a27fd6f14-kube-api-access-bmrpc\") pod \"packageserver-d55dfcdfc-7kwbt\" (UID: \"f57a4a48-39f0-44f3-bbf8-982a27fd6f14\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103914 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4zg7\" (UniqueName: \"kubernetes.io/projected/4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121-kube-api-access-d4zg7\") pod \"package-server-manager-789f6589d5-nhsgq\" (UID: \"4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103941 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d43c1374-e913-423e-ab78-6e14b1c16bfd-machine-approver-tls\") pod \"machine-approver-56656f9798-hjsr5\" (UID: \"d43c1374-e913-423e-ab78-6e14b1c16bfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103960 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f57a4a48-39f0-44f3-bbf8-982a27fd6f14-webhook-cert\") pod \"packageserver-d55dfcdfc-7kwbt\" (UID: \"f57a4a48-39f0-44f3-bbf8-982a27fd6f14\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.103982 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6c8c2a1f-4652-4791-a114-ebb9a46659ec-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-q5lx6\" (UID: \"6c8c2a1f-4652-4791-a114-ebb9a46659ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.105347 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfffde2b-573c-418b-903b-9117e958dabc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mncr8\" (UID: \"dfffde2b-573c-418b-903b-9117e958dabc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.105611 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-config\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.105881 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f57a4a48-39f0-44f3-bbf8-982a27fd6f14-tmpfs\") pod \"packageserver-d55dfcdfc-7kwbt\" (UID: \"f57a4a48-39f0-44f3-bbf8-982a27fd6f14\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.106040 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-m6p2k" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.106359 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-serving-cert\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.106387 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.106403 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f57a4a48-39f0-44f3-bbf8-982a27fd6f14-tmpfs\") pod \"packageserver-d55dfcdfc-7kwbt\" (UID: \"f57a4a48-39f0-44f3-bbf8-982a27fd6f14\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.107009 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d43c1374-e913-423e-ab78-6e14b1c16bfd-machine-approver-tls\") pod \"machine-approver-56656f9798-hjsr5\" (UID: \"d43c1374-e913-423e-ab78-6e14b1c16bfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.107302 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.107615 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d3bb5-c13d-49fc-9d5c-5b2ce0847d92-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4q2wd\" (UID: \"210d3bb5-c13d-49fc-9d5c-5b2ce0847d92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.107828 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-x829v"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.109148 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f57a4a48-39f0-44f3-bbf8-982a27fd6f14-webhook-cert\") pod \"packageserver-d55dfcdfc-7kwbt\" (UID: \"f57a4a48-39f0-44f3-bbf8-982a27fd6f14\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.109743 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e923a2-9b5c-49db-8a02-5eff8ca31033-serving-cert\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.110064 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.110408 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-config\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.112624 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-whmt5"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.113291 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.113823 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/402543c2-2831-4d3b-87ca-6d284c6d9be1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-h58tr\" (UID: \"402543c2-2831-4d3b-87ca-6d284c6d9be1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.114029 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.114720 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.115842 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.116878 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.118165 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dnmb8"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.119550 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.122928 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lcx97"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.126858 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.135046 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.136634 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-s6qsc"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.137938 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.140660 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nz8lf"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.146787 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.148153 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7b762"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.149265 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.151220 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dxl4k"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.153555 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-lmz7d"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.155187 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.155961 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-d6n9s"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.156980 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.158354 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.159201 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.160739 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8g9pk"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.161953 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.163426 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-857s6"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.166233 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-d6n9s"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.166650 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.169022 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-ccgf6"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.169776 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ccgf6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.173881 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-r6dwq"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.175215 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-h695r"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.175240 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-796c7"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.175346 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.185839 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-r6dwq"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.190219 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.206431 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.209845 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.209892 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2c7da228-335d-4f50-83e6-4ae152ed4874-console-oauth-config\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210639 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39580550-6421-4eb7-8727-42c26ad5c3f2-config\") pod \"route-controller-manager-6576b87f9c-f7s66\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210671 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85chc\" (UniqueName: \"kubernetes.io/projected/39580550-6421-4eb7-8727-42c26ad5c3f2-kube-api-access-85chc\") pod \"route-controller-manager-6576b87f9c-f7s66\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210693 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/29f23275-ee4d-41d6-acaa-a9c1363d71c3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p58wk\" (UID: \"29f23275-ee4d-41d6-acaa-a9c1363d71c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210713 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bdb30229-c0c5-4f40-a69c-6d8548cb76d8-profile-collector-cert\") pod \"catalog-operator-68c6474976-cx7j6\" (UID: \"bdb30229-c0c5-4f40-a69c-6d8548cb76d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210731 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1c808375-740d-4fe5-ae91-64f776b28e8a-stats-auth\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210773 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-etcd-serving-ca\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210792 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210822 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfffde2b-573c-418b-903b-9117e958dabc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mncr8\" (UID: \"dfffde2b-573c-418b-903b-9117e958dabc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210838 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-node-pullsecrets\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210853 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/29f23275-ee4d-41d6-acaa-a9c1363d71c3-srv-cert\") pod \"olm-operator-6b444d44fb-p58wk\" (UID: \"29f23275-ee4d-41d6-acaa-a9c1363d71c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210869 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df2ba163-1935-4600-a6c8-dd73d151ed32-secret-volume\") pod \"collect-profiles-29523795-hd6q4\" (UID: \"df2ba163-1935-4600-a6c8-dd73d151ed32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210886 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c7da228-335d-4f50-83e6-4ae152ed4874-console-serving-cert\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210958 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-audit\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210979 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.210998 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zrzs\" (UniqueName: \"kubernetes.io/projected/bdb30229-c0c5-4f40-a69c-6d8548cb76d8-kube-api-access-2zrzs\") pod \"catalog-operator-68c6474976-cx7j6\" (UID: \"bdb30229-c0c5-4f40-a69c-6d8548cb76d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211017 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211036 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/605d503b-946e-47e6-b08f-c6eaf0c3eb9a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-x9mtt\" (UID: \"605d503b-946e-47e6-b08f-c6eaf0c3eb9a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211057 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djw7b\" (UniqueName: \"kubernetes.io/projected/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-kube-api-access-djw7b\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211075 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c808375-740d-4fe5-ae91-64f776b28e8a-service-ca-bundle\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211091 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df2ba163-1935-4600-a6c8-dd73d151ed32-config-volume\") pod \"collect-profiles-29523795-hd6q4\" (UID: \"df2ba163-1935-4600-a6c8-dd73d151ed32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211107 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211126 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-trusted-ca-bundle\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211145 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/954a7c44-9845-4a84-8cd6-a5fe9be2d7d2-proxy-tls\") pod \"machine-config-controller-84d6567774-x829v\" (UID: \"954a7c44-9845-4a84-8cd6-a5fe9be2d7d2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211176 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1c808375-740d-4fe5-ae91-64f776b28e8a-default-certificate\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211227 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211249 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-969jw\" (UniqueName: \"kubernetes.io/projected/57795f27-c9c3-4f38-a5da-364e2488f343-kube-api-access-969jw\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211269 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmmjt\" (UniqueName: \"kubernetes.io/projected/5985084a-4c95-41f7-a40a-50550738e31c-kube-api-access-fmmjt\") pod \"downloads-7954f5f757-td7d6\" (UID: \"5985084a-4c95-41f7-a40a-50550738e31c\") " pod="openshift-console/downloads-7954f5f757-td7d6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211289 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-trusted-ca-bundle\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211282 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39580550-6421-4eb7-8727-42c26ad5c3f2-config\") pod \"route-controller-manager-6576b87f9c-f7s66\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211310 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6c8c2a1f-4652-4791-a114-ebb9a46659ec-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-q5lx6\" (UID: \"6c8c2a1f-4652-4791-a114-ebb9a46659ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211330 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfffde2b-573c-418b-903b-9117e958dabc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mncr8\" (UID: \"dfffde2b-573c-418b-903b-9117e958dabc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211353 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-image-import-ca\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211372 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-encryption-config\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211390 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtghp\" (UniqueName: \"kubernetes.io/projected/954a7c44-9845-4a84-8cd6-a5fe9be2d7d2-kube-api-access-jtghp\") pod \"machine-config-controller-84d6567774-x829v\" (UID: \"954a7c44-9845-4a84-8cd6-a5fe9be2d7d2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211414 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39580550-6421-4eb7-8727-42c26ad5c3f2-serving-cert\") pod \"route-controller-manager-6576b87f9c-f7s66\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211433 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th4hp\" (UniqueName: \"kubernetes.io/projected/ba1be0b0-ffa7-4b12-bc99-354621c8dd5e-kube-api-access-th4hp\") pod \"multus-admission-controller-857f4d67dd-nz8lf\" (UID: \"ba1be0b0-ffa7-4b12-bc99-354621c8dd5e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nz8lf" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211454 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1c808375-740d-4fe5-ae91-64f776b28e8a-metrics-certs\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211474 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211498 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-oauth-serving-cert\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211544 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e058a32b-9d0b-449a-a925-dc19858e6546-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bdhng\" (UID: \"e058a32b-9d0b-449a-a925-dc19858e6546\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211563 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211582 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-audit-dir\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211598 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-etcd-client\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211612 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-console-config\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211635 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211654 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjnc7\" (UniqueName: \"kubernetes.io/projected/605d503b-946e-47e6-b08f-c6eaf0c3eb9a-kube-api-access-pjnc7\") pod \"control-plane-machine-set-operator-78cbb6b69f-x9mtt\" (UID: \"605d503b-946e-47e6-b08f-c6eaf0c3eb9a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211688 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-serving-cert\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211711 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ba1be0b0-ffa7-4b12-bc99-354621c8dd5e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nz8lf\" (UID: \"ba1be0b0-ffa7-4b12-bc99-354621c8dd5e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nz8lf" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211731 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dnmb8\" (UID: \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211800 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-audit-policies\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211820 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57795f27-c9c3-4f38-a5da-364e2488f343-audit-dir\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211848 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w7n5\" (UniqueName: \"kubernetes.io/projected/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-kube-api-access-9w7n5\") pod \"marketplace-operator-79b997595-dnmb8\" (UID: \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211883 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxx4s\" (UniqueName: \"kubernetes.io/projected/2c7da228-335d-4f50-83e6-4ae152ed4874-kube-api-access-dxx4s\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211900 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211909 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-node-pullsecrets\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.211924 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76sj5\" (UniqueName: \"kubernetes.io/projected/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-kube-api-access-76sj5\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212013 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrwtn\" (UniqueName: \"kubernetes.io/projected/df2ba163-1935-4600-a6c8-dd73d151ed32-kube-api-access-lrwtn\") pod \"collect-profiles-29523795-hd6q4\" (UID: \"df2ba163-1935-4600-a6c8-dd73d151ed32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212038 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39580550-6421-4eb7-8727-42c26ad5c3f2-client-ca\") pod \"route-controller-manager-6576b87f9c-f7s66\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212063 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvqb9\" (UniqueName: \"kubernetes.io/projected/e058a32b-9d0b-449a-a925-dc19858e6546-kube-api-access-nvqb9\") pod \"openshift-apiserver-operator-796bbdcf4f-bdhng\" (UID: \"e058a32b-9d0b-449a-a925-dc19858e6546\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212099 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/954a7c44-9845-4a84-8cd6-a5fe9be2d7d2-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-x829v\" (UID: \"954a7c44-9845-4a84-8cd6-a5fe9be2d7d2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212124 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrwvb\" (UniqueName: \"kubernetes.io/projected/29f23275-ee4d-41d6-acaa-a9c1363d71c3-kube-api-access-lrwvb\") pod \"olm-operator-6b444d44fb-p58wk\" (UID: \"29f23275-ee4d-41d6-acaa-a9c1363d71c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212149 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntt9n\" (UniqueName: \"kubernetes.io/projected/1c808375-740d-4fe5-ae91-64f776b28e8a-kube-api-access-ntt9n\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212166 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bdb30229-c0c5-4f40-a69c-6d8548cb76d8-srv-cert\") pod \"catalog-operator-68c6474976-cx7j6\" (UID: \"bdb30229-c0c5-4f40-a69c-6d8548cb76d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212184 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-nhsgq\" (UID: \"4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212212 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4zg7\" (UniqueName: \"kubernetes.io/projected/4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121-kube-api-access-d4zg7\") pod \"package-server-manager-789f6589d5-nhsgq\" (UID: \"4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212237 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e058a32b-9d0b-449a-a925-dc19858e6546-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bdhng\" (UID: \"e058a32b-9d0b-449a-a925-dc19858e6546\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212265 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-etcd-serving-ca\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212291 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-service-ca\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212314 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfffde2b-573c-418b-903b-9117e958dabc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mncr8\" (UID: \"dfffde2b-573c-418b-903b-9117e958dabc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212335 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-config\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212353 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dggsh\" (UniqueName: \"kubernetes.io/projected/6c8c2a1f-4652-4791-a114-ebb9a46659ec-kube-api-access-dggsh\") pod \"cluster-samples-operator-665b6dd947-q5lx6\" (UID: \"6c8c2a1f-4652-4791-a114-ebb9a46659ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212374 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dnmb8\" (UID: \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212393 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.212478 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfffde2b-573c-418b-903b-9117e958dabc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mncr8\" (UID: \"dfffde2b-573c-418b-903b-9117e958dabc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.214132 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-audit\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.214798 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-trusted-ca-bundle\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.214962 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-console-config\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.215123 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/954a7c44-9845-4a84-8cd6-a5fe9be2d7d2-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-x829v\" (UID: \"954a7c44-9845-4a84-8cd6-a5fe9be2d7d2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.215112 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-trusted-ca-bundle\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.215902 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-config\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.215963 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-service-ca\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.216549 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bdb30229-c0c5-4f40-a69c-6d8548cb76d8-profile-collector-cert\") pod \"catalog-operator-68c6474976-cx7j6\" (UID: \"bdb30229-c0c5-4f40-a69c-6d8548cb76d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.216591 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-oauth-serving-cert\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.216617 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57795f27-c9c3-4f38-a5da-364e2488f343-audit-dir\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.216906 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-audit-dir\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.217908 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-image-import-ca\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.218094 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39580550-6421-4eb7-8727-42c26ad5c3f2-client-ca\") pod \"route-controller-manager-6576b87f9c-f7s66\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.218551 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c7da228-335d-4f50-83e6-4ae152ed4874-console-serving-cert\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.219025 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dnmb8\" (UID: \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.221137 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-serving-cert\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.221193 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/605d503b-946e-47e6-b08f-c6eaf0c3eb9a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-x9mtt\" (UID: \"605d503b-946e-47e6-b08f-c6eaf0c3eb9a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.221713 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2c7da228-335d-4f50-83e6-4ae152ed4874-console-oauth-config\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.222808 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-nhsgq\" (UID: \"4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.222857 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfffde2b-573c-418b-903b-9117e958dabc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mncr8\" (UID: \"dfffde2b-573c-418b-903b-9117e958dabc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.223504 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/29f23275-ee4d-41d6-acaa-a9c1363d71c3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p58wk\" (UID: \"29f23275-ee4d-41d6-acaa-a9c1363d71c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.223614 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39580550-6421-4eb7-8727-42c26ad5c3f2-serving-cert\") pod \"route-controller-manager-6576b87f9c-f7s66\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.224411 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bdb30229-c0c5-4f40-a69c-6d8548cb76d8-srv-cert\") pod \"catalog-operator-68c6474976-cx7j6\" (UID: \"bdb30229-c0c5-4f40-a69c-6d8548cb76d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.224521 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/29f23275-ee4d-41d6-acaa-a9c1363d71c3-srv-cert\") pod \"olm-operator-6b444d44fb-p58wk\" (UID: \"29f23275-ee4d-41d6-acaa-a9c1363d71c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.225183 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ba1be0b0-ffa7-4b12-bc99-354621c8dd5e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nz8lf\" (UID: \"ba1be0b0-ffa7-4b12-bc99-354621c8dd5e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nz8lf" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.225579 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6c8c2a1f-4652-4791-a114-ebb9a46659ec-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-q5lx6\" (UID: \"6c8c2a1f-4652-4791-a114-ebb9a46659ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.225827 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.226172 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df2ba163-1935-4600-a6c8-dd73d151ed32-secret-volume\") pod \"collect-profiles-29523795-hd6q4\" (UID: \"df2ba163-1935-4600-a6c8-dd73d151ed32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.227016 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-etcd-client\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.230337 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-encryption-config\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.246021 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.268571 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.295817 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.306441 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.313434 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e058a32b-9d0b-449a-a925-dc19858e6546-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bdhng\" (UID: \"e058a32b-9d0b-449a-a925-dc19858e6546\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.313580 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvqb9\" (UniqueName: \"kubernetes.io/projected/e058a32b-9d0b-449a-a925-dc19858e6546-kube-api-access-nvqb9\") pod \"openshift-apiserver-operator-796bbdcf4f-bdhng\" (UID: \"e058a32b-9d0b-449a-a925-dc19858e6546\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.313651 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e058a32b-9d0b-449a-a925-dc19858e6546-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bdhng\" (UID: \"e058a32b-9d0b-449a-a925-dc19858e6546\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.314673 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e058a32b-9d0b-449a-a925-dc19858e6546-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bdhng\" (UID: \"e058a32b-9d0b-449a-a925-dc19858e6546\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.317326 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e058a32b-9d0b-449a-a925-dc19858e6546-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bdhng\" (UID: \"e058a32b-9d0b-449a-a925-dc19858e6546\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.323540 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dnmb8\" (UID: \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.328263 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.343304 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-m6p2k"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.349323 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.368434 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.369170 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.375911 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df2ba163-1935-4600-a6c8-dd73d151ed32-config-volume\") pod \"collect-profiles-29523795-hd6q4\" (UID: \"df2ba163-1935-4600-a6c8-dd73d151ed32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.385946 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.413489 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.413932 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.425624 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 18 15:24:11 crc kubenswrapper[4968]: W0218 15:24:11.429332 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03d43fa6_7663_45ca_b028_0d245cc505af.slice/crio-9c2f32183241e5c97ca5095b327b118e853deb07e9ca1e5943e4a08a7ead5750 WatchSource:0}: Error finding container 9c2f32183241e5c97ca5095b327b118e853deb07e9ca1e5943e4a08a7ead5750: Status 404 returned error can't find the container with id 9c2f32183241e5c97ca5095b327b118e853deb07e9ca1e5943e4a08a7ead5750 Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.436690 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x"] Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.446542 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.448962 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-audit-policies\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.466879 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.485960 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.506172 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.525737 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.546469 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.566523 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.586318 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.601363 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.606278 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.615023 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.626029 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.639664 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.646266 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.666335 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.670243 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.687793 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.697291 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.706758 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.712998 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.753656 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.756480 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.765039 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.765657 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.767804 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.775483 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.786009 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.810210 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.815703 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.825953 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.845782 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.848232 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.869395 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.877118 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/954a7c44-9845-4a84-8cd6-a5fe9be2d7d2-proxy-tls\") pod \"machine-config-controller-84d6567774-x829v\" (UID: \"954a7c44-9845-4a84-8cd6-a5fe9be2d7d2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.885819 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.926404 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.946822 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.966510 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 18 15:24:11 crc kubenswrapper[4968]: I0218 15:24:11.986120 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.006132 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.015496 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c808375-740d-4fe5-ae91-64f776b28e8a-service-ca-bundle\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.025929 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.037144 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1c808375-740d-4fe5-ae91-64f776b28e8a-stats-auth\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.043964 4968 request.go:700] Waited for 1.008027443s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/secrets?fieldSelector=metadata.name%3Drouter-metrics-certs-default&limit=500&resourceVersion=0 Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.046283 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.051033 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1c808375-740d-4fe5-ae91-64f776b28e8a-metrics-certs\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.068074 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.088911 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.093350 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" event={"ID":"03d43fa6-7663-45ca-b028-0d245cc505af","Type":"ContainerStarted","Data":"483e984fd2024ec0040b6ed1e9416b85a8be62db11847f8fc4be22c548708bd7"} Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.093556 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" event={"ID":"03d43fa6-7663-45ca-b028-0d245cc505af","Type":"ContainerStarted","Data":"9c2f32183241e5c97ca5095b327b118e853deb07e9ca1e5943e4a08a7ead5750"} Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.095667 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-m6p2k" event={"ID":"a8b4bf38-6d44-4b8a-8c08-afa0023e6d28","Type":"ContainerStarted","Data":"46565b1bdbcc8e5024187f7d319b778f2f7be68d208711489b7eff5bc4fa66c8"} Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.095727 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-m6p2k" event={"ID":"a8b4bf38-6d44-4b8a-8c08-afa0023e6d28","Type":"ContainerStarted","Data":"ea250f10360332574b425f88178836980eea3651c46a5f28bb4695b44ce7a12a"} Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.095772 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-m6p2k" event={"ID":"a8b4bf38-6d44-4b8a-8c08-afa0023e6d28","Type":"ContainerStarted","Data":"c74456b8dcd5b7f45d172ae1543232c45ea09a04c39c433cdc84ccf0251a075c"} Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.096810 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1c808375-740d-4fe5-ae91-64f776b28e8a-default-certificate\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.097366 4968 generic.go:334] "Generic (PLEG): container finished" podID="5ecce765-efd2-479b-9512-295892b014a7" containerID="1a6565c19dfa565c4ad3d62f485531d99a2976496335c817c3b08732d59481d8" exitCode=0 Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.097435 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" event={"ID":"5ecce765-efd2-479b-9512-295892b014a7","Type":"ContainerDied","Data":"1a6565c19dfa565c4ad3d62f485531d99a2976496335c817c3b08732d59481d8"} Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.097456 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" event={"ID":"5ecce765-efd2-479b-9512-295892b014a7","Type":"ContainerStarted","Data":"2ea5750fdd9af4ac8a969faab44d1f29ab041e2075032fded47d9731c4fcbb50"} Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.098890 4968 generic.go:334] "Generic (PLEG): container finished" podID="925dd643-0943-43dd-a71e-c7cbcfae2c00" containerID="94504490dc1638b37df54772b389c2e8b8e8118f0b0f35890867e09483348ad2" exitCode=0 Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.098943 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" event={"ID":"925dd643-0943-43dd-a71e-c7cbcfae2c00","Type":"ContainerDied","Data":"94504490dc1638b37df54772b389c2e8b8e8118f0b0f35890867e09483348ad2"} Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.099243 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" event={"ID":"925dd643-0943-43dd-a71e-c7cbcfae2c00","Type":"ContainerStarted","Data":"804e3fb90cee10778a2e26aa582053d26d2a4ab1e0aa4c618f8a3faf1b41e865"} Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.106593 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.127241 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.146107 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.167908 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.185896 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.209310 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.226598 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.246555 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.265998 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.286000 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.312467 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.326406 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.346635 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.367917 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.386252 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.405973 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.425951 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.446413 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.465561 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.486901 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.512936 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.546255 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.567434 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.586033 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.607800 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.642085 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz4k2\" (UniqueName: \"kubernetes.io/projected/41a0f2d1-405f-4396-a081-b874a35bacf8-kube-api-access-dz4k2\") pod \"console-operator-58897d9998-nkbhx\" (UID: \"41a0f2d1-405f-4396-a081-b874a35bacf8\") " pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.664163 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vljcg\" (UniqueName: \"kubernetes.io/projected/d8e923a2-9b5c-49db-8a02-5eff8ca31033-kube-api-access-vljcg\") pod \"controller-manager-879f6c89f-7b762\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.666172 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.700978 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv8hk\" (UniqueName: \"kubernetes.io/projected/402543c2-2831-4d3b-87ca-6d284c6d9be1-kube-api-access-xv8hk\") pod \"machine-api-operator-5694c8668f-h58tr\" (UID: \"402543c2-2831-4d3b-87ca-6d284c6d9be1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.709157 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.724308 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.725932 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.746485 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.766981 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.810425 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvf2k\" (UniqueName: \"kubernetes.io/projected/d43c1374-e913-423e-ab78-6e14b1c16bfd-kube-api-access-nvf2k\") pod \"machine-approver-56656f9798-hjsr5\" (UID: \"d43c1374-e913-423e-ab78-6e14b1c16bfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.826868 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qh7t\" (UniqueName: \"kubernetes.io/projected/210d3bb5-c13d-49fc-9d5c-5b2ce0847d92-kube-api-access-5qh7t\") pod \"openshift-controller-manager-operator-756b6f6bc6-4q2wd\" (UID: \"210d3bb5-c13d-49fc-9d5c-5b2ce0847d92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.842617 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmrpc\" (UniqueName: \"kubernetes.io/projected/f57a4a48-39f0-44f3-bbf8-982a27fd6f14-kube-api-access-bmrpc\") pod \"packageserver-d55dfcdfc-7kwbt\" (UID: \"f57a4a48-39f0-44f3-bbf8-982a27fd6f14\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.846827 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.865795 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.865862 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.877052 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.888156 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.889699 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7b762"] Feb 18 15:24:12 crc kubenswrapper[4968]: W0218 15:24:12.898217 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8e923a2_9b5c_49db_8a02_5eff8ca31033.slice/crio-5d78c8879c2bd5e1db2b40acf62fe01315492fc86172847957b9d2ae87475c06 WatchSource:0}: Error finding container 5d78c8879c2bd5e1db2b40acf62fe01315492fc86172847957b9d2ae87475c06: Status 404 returned error can't find the container with id 5d78c8879c2bd5e1db2b40acf62fe01315492fc86172847957b9d2ae87475c06 Feb 18 15:24:12 crc kubenswrapper[4968]: W0218 15:24:12.899628 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd43c1374_e913_423e_ab78_6e14b1c16bfd.slice/crio-a81e63e3183b9b6bc20b6b314f30242a3aa907101eb5463347ddf51c767a79b7 WatchSource:0}: Error finding container a81e63e3183b9b6bc20b6b314f30242a3aa907101eb5463347ddf51c767a79b7: Status 404 returned error can't find the container with id a81e63e3183b9b6bc20b6b314f30242a3aa907101eb5463347ddf51c767a79b7 Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.905693 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.914657 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.928909 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.947279 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.966537 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.988082 4968 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 18 15:24:12 crc kubenswrapper[4968]: I0218 15:24:12.990113 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.006234 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.044318 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76sj5\" (UniqueName: \"kubernetes.io/projected/5814e0fe-15f6-4675-ab06-ca9f9a467cd0-kube-api-access-76sj5\") pod \"etcd-operator-b45778765-6v575\" (UID: \"5814e0fe-15f6-4675-ab06-ca9f9a467cd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.044436 4968 request.go:700] Waited for 1.831730256s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/serviceaccounts/route-controller-manager-sa/token Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.069241 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85chc\" (UniqueName: \"kubernetes.io/projected/39580550-6421-4eb7-8727-42c26ad5c3f2-kube-api-access-85chc\") pod \"route-controller-manager-6576b87f9c-f7s66\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.071534 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nkbhx"] Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.086854 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrwtn\" (UniqueName: \"kubernetes.io/projected/df2ba163-1935-4600-a6c8-dd73d151ed32-kube-api-access-lrwtn\") pod \"collect-profiles-29523795-hd6q4\" (UID: \"df2ba163-1935-4600-a6c8-dd73d151ed32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.092126 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.105427 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.115123 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" event={"ID":"d43c1374-e913-423e-ab78-6e14b1c16bfd","Type":"ContainerStarted","Data":"a81e63e3183b9b6bc20b6b314f30242a3aa907101eb5463347ddf51c767a79b7"} Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.116327 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" event={"ID":"d8e923a2-9b5c-49db-8a02-5eff8ca31033","Type":"ContainerStarted","Data":"3d4acda592e3e185edc66c28d123e85a21d9d5e4ec6ee31c4c9304483298435d"} Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.116380 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" event={"ID":"d8e923a2-9b5c-49db-8a02-5eff8ca31033","Type":"ContainerStarted","Data":"5d78c8879c2bd5e1db2b40acf62fe01315492fc86172847957b9d2ae87475c06"} Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.117516 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.119320 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4zg7\" (UniqueName: \"kubernetes.io/projected/4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121-kube-api-access-d4zg7\") pod \"package-server-manager-789f6589d5-nhsgq\" (UID: \"4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.123945 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" event={"ID":"925dd643-0943-43dd-a71e-c7cbcfae2c00","Type":"ContainerStarted","Data":"9b97a485941cf9cf02352b27f827ca055fb3cb42cff4b7b7b7dc08c11fe28954"} Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.124992 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.125183 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtghp\" (UniqueName: \"kubernetes.io/projected/954a7c44-9845-4a84-8cd6-a5fe9be2d7d2-kube-api-access-jtghp\") pod \"machine-config-controller-84d6567774-x829v\" (UID: \"954a7c44-9845-4a84-8cd6-a5fe9be2d7d2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.125217 4968 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-7b762 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.125262 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" podUID="d8e923a2-9b5c-49db-8a02-5eff8ca31033" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.126647 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-nkbhx" event={"ID":"41a0f2d1-405f-4396-a081-b874a35bacf8","Type":"ContainerStarted","Data":"97614b8cc449535d8e41d9598e736aaf01dbd8ae9f048c3c5aff52b167c29071"} Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.133408 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" event={"ID":"5ecce765-efd2-479b-9512-295892b014a7","Type":"ContainerStarted","Data":"9b387150ec5c54470dc893bc8cad22d40a07fca82e66c7d5b236ef462dd6bb22"} Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.146450 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-969jw\" (UniqueName: \"kubernetes.io/projected/57795f27-c9c3-4f38-a5da-364e2488f343-kube-api-access-969jw\") pod \"oauth-openshift-558db77b4-8g9pk\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.175962 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmmjt\" (UniqueName: \"kubernetes.io/projected/5985084a-4c95-41f7-a40a-50550738e31c-kube-api-access-fmmjt\") pod \"downloads-7954f5f757-td7d6\" (UID: \"5985084a-4c95-41f7-a40a-50550738e31c\") " pod="openshift-console/downloads-7954f5f757-td7d6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.192069 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrwvb\" (UniqueName: \"kubernetes.io/projected/29f23275-ee4d-41d6-acaa-a9c1363d71c3-kube-api-access-lrwvb\") pod \"olm-operator-6b444d44fb-p58wk\" (UID: \"29f23275-ee4d-41d6-acaa-a9c1363d71c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.206643 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntt9n\" (UniqueName: \"kubernetes.io/projected/1c808375-740d-4fe5-ae91-64f776b28e8a-kube-api-access-ntt9n\") pod \"router-default-5444994796-w9zh9\" (UID: \"1c808375-740d-4fe5-ae91-64f776b28e8a\") " pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.208261 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.222410 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djw7b\" (UniqueName: \"kubernetes.io/projected/f9194c82-960a-4dea-94ef-5b2d5b4e4a75-kube-api-access-djw7b\") pod \"apiserver-76f77b778f-whmt5\" (UID: \"f9194c82-960a-4dea-94ef-5b2d5b4e4a75\") " pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.232025 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.232644 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.241510 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt"] Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.245464 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zrzs\" (UniqueName: \"kubernetes.io/projected/bdb30229-c0c5-4f40-a69c-6d8548cb76d8-kube-api-access-2zrzs\") pod \"catalog-operator-68c6474976-cx7j6\" (UID: \"bdb30229-c0c5-4f40-a69c-6d8548cb76d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:13 crc kubenswrapper[4968]: W0218 15:24:13.254936 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf57a4a48_39f0_44f3_bbf8_982a27fd6f14.slice/crio-1ee8443ee3a0b43dcf64bd0757eb068c98403c66acaa92812a17974ed4fd8489 WatchSource:0}: Error finding container 1ee8443ee3a0b43dcf64bd0757eb068c98403c66acaa92812a17974ed4fd8489: Status 404 returned error can't find the container with id 1ee8443ee3a0b43dcf64bd0757eb068c98403c66acaa92812a17974ed4fd8489 Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.259046 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.266587 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.268452 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjnc7\" (UniqueName: \"kubernetes.io/projected/605d503b-946e-47e6-b08f-c6eaf0c3eb9a-kube-api-access-pjnc7\") pod \"control-plane-machine-set-operator-78cbb6b69f-x9mtt\" (UID: \"605d503b-946e-47e6-b08f-c6eaf0c3eb9a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.287357 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.292194 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfffde2b-573c-418b-903b-9117e958dabc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mncr8\" (UID: \"dfffde2b-573c-418b-903b-9117e958dabc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.304779 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-h58tr"] Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.316409 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dggsh\" (UniqueName: \"kubernetes.io/projected/6c8c2a1f-4652-4791-a114-ebb9a46659ec-kube-api-access-dggsh\") pod \"cluster-samples-operator-665b6dd947-q5lx6\" (UID: \"6c8c2a1f-4652-4791-a114-ebb9a46659ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.325234 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th4hp\" (UniqueName: \"kubernetes.io/projected/ba1be0b0-ffa7-4b12-bc99-354621c8dd5e-kube-api-access-th4hp\") pod \"multus-admission-controller-857f4d67dd-nz8lf\" (UID: \"ba1be0b0-ffa7-4b12-bc99-354621c8dd5e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nz8lf" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.350060 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxx4s\" (UniqueName: \"kubernetes.io/projected/2c7da228-335d-4f50-83e6-4ae152ed4874-kube-api-access-dxx4s\") pod \"console-f9d7485db-s6qsc\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.363031 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w7n5\" (UniqueName: \"kubernetes.io/projected/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-kube-api-access-9w7n5\") pod \"marketplace-operator-79b997595-dnmb8\" (UID: \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.375418 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.388684 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvqb9\" (UniqueName: \"kubernetes.io/projected/e058a32b-9d0b-449a-a925-dc19858e6546-kube-api-access-nvqb9\") pod \"openshift-apiserver-operator-796bbdcf4f-bdhng\" (UID: \"e058a32b-9d0b-449a-a925-dc19858e6546\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.421970 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.429201 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.433723 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.442635 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-td7d6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447343 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f1b7574f-d70f-46dd-ac82-2cc55659bf7d-metrics-tls\") pod \"ingress-operator-5b745b69d9-pqd7g\" (UID: \"f1b7574f-d70f-46dd-ac82-2cc55659bf7d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447388 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/46d0135b-0668-4024-aaee-8f28eb43f235-images\") pod \"machine-config-operator-74547568cd-k64ds\" (UID: \"46d0135b-0668-4024-aaee-8f28eb43f235\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447410 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac98f4f8-45d4-4bce-b635-8be96ba93afd-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-4sklv\" (UID: \"ac98f4f8-45d4-4bce-b635-8be96ba93afd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447434 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-trusted-ca\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447449 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-bound-sa-token\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447467 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e122baf6-f5c4-4f82-87ee-520e14095f73-signing-cabundle\") pod \"service-ca-9c57cc56f-dxl4k\" (UID: \"e122baf6-f5c4-4f82-87ee-520e14095f73\") " pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447491 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5-config\") pod \"kube-apiserver-operator-766d6c64bb-kv8lj\" (UID: \"2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447508 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pd85\" (UniqueName: \"kubernetes.io/projected/f1b7574f-d70f-46dd-ac82-2cc55659bf7d-kube-api-access-2pd85\") pod \"ingress-operator-5b745b69d9-pqd7g\" (UID: \"f1b7574f-d70f-46dd-ac82-2cc55659bf7d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447534 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-installation-pull-secrets\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447562 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kv8lj\" (UID: \"2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447577 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e122baf6-f5c4-4f82-87ee-520e14095f73-signing-key\") pod \"service-ca-9c57cc56f-dxl4k\" (UID: \"e122baf6-f5c4-4f82-87ee-520e14095f73\") " pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447594 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfrmc\" (UniqueName: \"kubernetes.io/projected/46d0135b-0668-4024-aaee-8f28eb43f235-kube-api-access-wfrmc\") pod \"machine-config-operator-74547568cd-k64ds\" (UID: \"46d0135b-0668-4024-aaee-8f28eb43f235\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447611 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb43c877-a415-4362-a629-8b77f5bce23f-serving-cert\") pod \"service-ca-operator-777779d784-857s6\" (UID: \"fb43c877-a415-4362-a629-8b77f5bce23f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447651 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-registry-certificates\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447673 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac98f4f8-45d4-4bce-b635-8be96ba93afd-config\") pod \"kube-controller-manager-operator-78b949d7b-4sklv\" (UID: \"ac98f4f8-45d4-4bce-b635-8be96ba93afd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447711 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447729 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f1b7574f-d70f-46dd-ac82-2cc55659bf7d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pqd7g\" (UID: \"f1b7574f-d70f-46dd-ac82-2cc55659bf7d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447800 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1b7574f-d70f-46dd-ac82-2cc55659bf7d-trusted-ca\") pod \"ingress-operator-5b745b69d9-pqd7g\" (UID: \"f1b7574f-d70f-46dd-ac82-2cc55659bf7d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447819 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac98f4f8-45d4-4bce-b635-8be96ba93afd-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4sklv\" (UID: \"ac98f4f8-45d4-4bce-b635-8be96ba93afd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447838 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsph7\" (UniqueName: \"kubernetes.io/projected/e122baf6-f5c4-4f82-87ee-520e14095f73-kube-api-access-nsph7\") pod \"service-ca-9c57cc56f-dxl4k\" (UID: \"e122baf6-f5c4-4f82-87ee-520e14095f73\") " pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447857 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/46d0135b-0668-4024-aaee-8f28eb43f235-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k64ds\" (UID: \"46d0135b-0668-4024-aaee-8f28eb43f235\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.447874 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-ca-trust-extracted\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.448013 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb43c877-a415-4362-a629-8b77f5bce23f-config\") pod \"service-ca-operator-777779d784-857s6\" (UID: \"fb43c877-a415-4362-a629-8b77f5bce23f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.448163 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kv8lj\" (UID: \"2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.448198 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwjjb\" (UniqueName: \"kubernetes.io/projected/fb43c877-a415-4362-a629-8b77f5bce23f-kube-api-access-kwjjb\") pod \"service-ca-operator-777779d784-857s6\" (UID: \"fb43c877-a415-4362-a629-8b77f5bce23f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.448230 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-registry-tls\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.448256 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/46d0135b-0668-4024-aaee-8f28eb43f235-proxy-tls\") pod \"machine-config-operator-74547568cd-k64ds\" (UID: \"46d0135b-0668-4024-aaee-8f28eb43f235\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.448290 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwqbc\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-kube-api-access-kwqbc\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: E0218 15:24:13.448394 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:13.948375261 +0000 UTC m=+153.333820123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.450033 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.498079 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-nz8lf" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.526052 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.535628 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.559397 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.559665 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e122baf6-f5c4-4f82-87ee-520e14095f73-signing-key\") pod \"service-ca-9c57cc56f-dxl4k\" (UID: \"e122baf6-f5c4-4f82-87ee-520e14095f73\") " pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.559707 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqqrd\" (UniqueName: \"kubernetes.io/projected/93887080-7c19-4f58-bc8d-9eff7323a02e-kube-api-access-xqqrd\") pod \"dns-default-d6n9s\" (UID: \"93887080-7c19-4f58-bc8d-9eff7323a02e\") " pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.559764 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfrmc\" (UniqueName: \"kubernetes.io/projected/46d0135b-0668-4024-aaee-8f28eb43f235-kube-api-access-wfrmc\") pod \"machine-config-operator-74547568cd-k64ds\" (UID: \"46d0135b-0668-4024-aaee-8f28eb43f235\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.559802 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10ca57c1-d445-4335-8051-f90489bdcacf-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hmnt5\" (UID: \"10ca57c1-d445-4335-8051-f90489bdcacf\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.559823 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb43c877-a415-4362-a629-8b77f5bce23f-serving-cert\") pod \"service-ca-operator-777779d784-857s6\" (UID: \"fb43c877-a415-4362-a629-8b77f5bce23f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.559845 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4692355-bdb5-4904-86f9-7b01f41aded7-serving-cert\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.559865 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-registration-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.559932 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-registry-certificates\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.559954 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-mountpoint-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.559984 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-plugins-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.560027 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4692355-bdb5-4904-86f9-7b01f41aded7-service-ca-bundle\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.560068 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac98f4f8-45d4-4bce-b635-8be96ba93afd-config\") pod \"kube-controller-manager-operator-78b949d7b-4sklv\" (UID: \"ac98f4f8-45d4-4bce-b635-8be96ba93afd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564044 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-socket-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564150 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d212b310-5c16-4fdf-9804-39eb782d6478-certs\") pod \"machine-config-server-ccgf6\" (UID: \"d212b310-5c16-4fdf-9804-39eb782d6478\") " pod="openshift-machine-config-operator/machine-config-server-ccgf6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564192 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f1b7574f-d70f-46dd-ac82-2cc55659bf7d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pqd7g\" (UID: \"f1b7574f-d70f-46dd-ac82-2cc55659bf7d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564242 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d90a8f89-0b24-47b4-bba6-2b84b7443f78-cert\") pod \"ingress-canary-h695r\" (UID: \"d90a8f89-0b24-47b4-bba6-2b84b7443f78\") " pod="openshift-ingress-canary/ingress-canary-h695r" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564320 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d212b310-5c16-4fdf-9804-39eb782d6478-node-bootstrap-token\") pod \"machine-config-server-ccgf6\" (UID: \"d212b310-5c16-4fdf-9804-39eb782d6478\") " pod="openshift-machine-config-operator/machine-config-server-ccgf6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564381 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1b7574f-d70f-46dd-ac82-2cc55659bf7d-trusted-ca\") pod \"ingress-operator-5b745b69d9-pqd7g\" (UID: \"f1b7574f-d70f-46dd-ac82-2cc55659bf7d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564403 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4692355-bdb5-4904-86f9-7b01f41aded7-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564426 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd2wc\" (UniqueName: \"kubernetes.io/projected/d90a8f89-0b24-47b4-bba6-2b84b7443f78-kube-api-access-bd2wc\") pod \"ingress-canary-h695r\" (UID: \"d90a8f89-0b24-47b4-bba6-2b84b7443f78\") " pod="openshift-ingress-canary/ingress-canary-h695r" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564448 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac98f4f8-45d4-4bce-b635-8be96ba93afd-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4sklv\" (UID: \"ac98f4f8-45d4-4bce-b635-8be96ba93afd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564470 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsph7\" (UniqueName: \"kubernetes.io/projected/e122baf6-f5c4-4f82-87ee-520e14095f73-kube-api-access-nsph7\") pod \"service-ca-9c57cc56f-dxl4k\" (UID: \"e122baf6-f5c4-4f82-87ee-520e14095f73\") " pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564493 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4692355-bdb5-4904-86f9-7b01f41aded7-config\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564558 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/46d0135b-0668-4024-aaee-8f28eb43f235-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k64ds\" (UID: \"46d0135b-0668-4024-aaee-8f28eb43f235\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564599 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-ca-trust-extracted\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564628 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/93887080-7c19-4f58-bc8d-9eff7323a02e-metrics-tls\") pod \"dns-default-d6n9s\" (UID: \"93887080-7c19-4f58-bc8d-9eff7323a02e\") " pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564650 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6sxw\" (UniqueName: \"kubernetes.io/projected/10ca57c1-d445-4335-8051-f90489bdcacf-kube-api-access-b6sxw\") pod \"kube-storage-version-migrator-operator-b67b599dd-hmnt5\" (UID: \"10ca57c1-d445-4335-8051-f90489bdcacf\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564775 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb43c877-a415-4362-a629-8b77f5bce23f-config\") pod \"service-ca-operator-777779d784-857s6\" (UID: \"fb43c877-a415-4362-a629-8b77f5bce23f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564896 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kv8lj\" (UID: \"2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564923 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwjjb\" (UniqueName: \"kubernetes.io/projected/fb43c877-a415-4362-a629-8b77f5bce23f-kube-api-access-kwjjb\") pod \"service-ca-operator-777779d784-857s6\" (UID: \"fb43c877-a415-4362-a629-8b77f5bce23f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564946 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-registry-tls\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.564971 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/46d0135b-0668-4024-aaee-8f28eb43f235-proxy-tls\") pod \"machine-config-operator-74547568cd-k64ds\" (UID: \"46d0135b-0668-4024-aaee-8f28eb43f235\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565056 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwqbc\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-kube-api-access-kwqbc\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565127 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2rmf\" (UniqueName: \"kubernetes.io/projected/f4692355-bdb5-4904-86f9-7b01f41aded7-kube-api-access-f2rmf\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565209 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f1b7574f-d70f-46dd-ac82-2cc55659bf7d-metrics-tls\") pod \"ingress-operator-5b745b69d9-pqd7g\" (UID: \"f1b7574f-d70f-46dd-ac82-2cc55659bf7d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565237 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgzps\" (UniqueName: \"kubernetes.io/projected/9a2c33ab-1bc7-436d-95b7-098dd2e57b40-kube-api-access-zgzps\") pod \"migrator-59844c95c7-796c7\" (UID: \"9a2c33ab-1bc7-436d-95b7-098dd2e57b40\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-796c7" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565259 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93887080-7c19-4f58-bc8d-9eff7323a02e-config-volume\") pod \"dns-default-d6n9s\" (UID: \"93887080-7c19-4f58-bc8d-9eff7323a02e\") " pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565300 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10ca57c1-d445-4335-8051-f90489bdcacf-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hmnt5\" (UID: \"10ca57c1-d445-4335-8051-f90489bdcacf\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565340 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/46d0135b-0668-4024-aaee-8f28eb43f235-images\") pod \"machine-config-operator-74547568cd-k64ds\" (UID: \"46d0135b-0668-4024-aaee-8f28eb43f235\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565366 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac98f4f8-45d4-4bce-b635-8be96ba93afd-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-4sklv\" (UID: \"ac98f4f8-45d4-4bce-b635-8be96ba93afd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565404 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-trusted-ca\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565458 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-bound-sa-token\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565481 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e122baf6-f5c4-4f82-87ee-520e14095f73-signing-cabundle\") pod \"service-ca-9c57cc56f-dxl4k\" (UID: \"e122baf6-f5c4-4f82-87ee-520e14095f73\") " pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565557 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5-config\") pod \"kube-apiserver-operator-766d6c64bb-kv8lj\" (UID: \"2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565584 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pd85\" (UniqueName: \"kubernetes.io/projected/f1b7574f-d70f-46dd-ac82-2cc55659bf7d-kube-api-access-2pd85\") pod \"ingress-operator-5b745b69d9-pqd7g\" (UID: \"f1b7574f-d70f-46dd-ac82-2cc55659bf7d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565624 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv822\" (UniqueName: \"kubernetes.io/projected/d212b310-5c16-4fdf-9804-39eb782d6478-kube-api-access-wv822\") pod \"machine-config-server-ccgf6\" (UID: \"d212b310-5c16-4fdf-9804-39eb782d6478\") " pod="openshift-machine-config-operator/machine-config-server-ccgf6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565720 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-installation-pull-secrets\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565770 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-csi-data-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565794 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf885\" (UniqueName: \"kubernetes.io/projected/c858a456-78e0-4f36-9a4e-6db413d94d48-kube-api-access-mf885\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.565949 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kv8lj\" (UID: \"2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" Feb 18 15:24:13 crc kubenswrapper[4968]: E0218 15:24:13.567477 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:14.067430745 +0000 UTC m=+153.452875617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.570552 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e122baf6-f5c4-4f82-87ee-520e14095f73-signing-cabundle\") pod \"service-ca-9c57cc56f-dxl4k\" (UID: \"e122baf6-f5c4-4f82-87ee-520e14095f73\") " pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.572071 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb43c877-a415-4362-a629-8b77f5bce23f-config\") pod \"service-ca-operator-777779d784-857s6\" (UID: \"fb43c877-a415-4362-a629-8b77f5bce23f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.572511 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-ca-trust-extracted\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.573369 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-registry-certificates\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.573977 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac98f4f8-45d4-4bce-b635-8be96ba93afd-config\") pod \"kube-controller-manager-operator-78b949d7b-4sklv\" (UID: \"ac98f4f8-45d4-4bce-b635-8be96ba93afd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.576066 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5-config\") pod \"kube-apiserver-operator-766d6c64bb-kv8lj\" (UID: \"2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.577585 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/46d0135b-0668-4024-aaee-8f28eb43f235-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k64ds\" (UID: \"46d0135b-0668-4024-aaee-8f28eb43f235\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.578512 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/46d0135b-0668-4024-aaee-8f28eb43f235-images\") pod \"machine-config-operator-74547568cd-k64ds\" (UID: \"46d0135b-0668-4024-aaee-8f28eb43f235\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.579970 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1b7574f-d70f-46dd-ac82-2cc55659bf7d-trusted-ca\") pod \"ingress-operator-5b745b69d9-pqd7g\" (UID: \"f1b7574f-d70f-46dd-ac82-2cc55659bf7d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.581343 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-trusted-ca\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.583903 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e122baf6-f5c4-4f82-87ee-520e14095f73-signing-key\") pod \"service-ca-9c57cc56f-dxl4k\" (UID: \"e122baf6-f5c4-4f82-87ee-520e14095f73\") " pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.586452 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfrmc\" (UniqueName: \"kubernetes.io/projected/46d0135b-0668-4024-aaee-8f28eb43f235-kube-api-access-wfrmc\") pod \"machine-config-operator-74547568cd-k64ds\" (UID: \"46d0135b-0668-4024-aaee-8f28eb43f235\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.587033 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-installation-pull-secrets\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.587644 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac98f4f8-45d4-4bce-b635-8be96ba93afd-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4sklv\" (UID: \"ac98f4f8-45d4-4bce-b635-8be96ba93afd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.587656 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/46d0135b-0668-4024-aaee-8f28eb43f235-proxy-tls\") pod \"machine-config-operator-74547568cd-k64ds\" (UID: \"46d0135b-0668-4024-aaee-8f28eb43f235\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.597941 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kv8lj\" (UID: \"2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.598362 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f1b7574f-d70f-46dd-ac82-2cc55659bf7d-metrics-tls\") pod \"ingress-operator-5b745b69d9-pqd7g\" (UID: \"f1b7574f-d70f-46dd-ac82-2cc55659bf7d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.598454 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb43c877-a415-4362-a629-8b77f5bce23f-serving-cert\") pod \"service-ca-operator-777779d784-857s6\" (UID: \"fb43c877-a415-4362-a629-8b77f5bce23f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.598647 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-registry-tls\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.632422 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f1b7574f-d70f-46dd-ac82-2cc55659bf7d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pqd7g\" (UID: \"f1b7574f-d70f-46dd-ac82-2cc55659bf7d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.645636 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.648292 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd"] Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.658109 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsph7\" (UniqueName: \"kubernetes.io/projected/e122baf6-f5c4-4f82-87ee-520e14095f73-kube-api-access-nsph7\") pod \"service-ca-9c57cc56f-dxl4k\" (UID: \"e122baf6-f5c4-4f82-87ee-520e14095f73\") " pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.666106 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.667813 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv822\" (UniqueName: \"kubernetes.io/projected/d212b310-5c16-4fdf-9804-39eb782d6478-kube-api-access-wv822\") pod \"machine-config-server-ccgf6\" (UID: \"d212b310-5c16-4fdf-9804-39eb782d6478\") " pod="openshift-machine-config-operator/machine-config-server-ccgf6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.667848 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-csi-data-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.667867 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf885\" (UniqueName: \"kubernetes.io/projected/c858a456-78e0-4f36-9a4e-6db413d94d48-kube-api-access-mf885\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.667893 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqqrd\" (UniqueName: \"kubernetes.io/projected/93887080-7c19-4f58-bc8d-9eff7323a02e-kube-api-access-xqqrd\") pod \"dns-default-d6n9s\" (UID: \"93887080-7c19-4f58-bc8d-9eff7323a02e\") " pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.667936 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10ca57c1-d445-4335-8051-f90489bdcacf-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hmnt5\" (UID: \"10ca57c1-d445-4335-8051-f90489bdcacf\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.667959 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4692355-bdb5-4904-86f9-7b01f41aded7-serving-cert\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.667974 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-registration-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.667993 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-mountpoint-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668010 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-plugins-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668027 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4692355-bdb5-4904-86f9-7b01f41aded7-service-ca-bundle\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668046 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-socket-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668066 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668082 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d212b310-5c16-4fdf-9804-39eb782d6478-certs\") pod \"machine-config-server-ccgf6\" (UID: \"d212b310-5c16-4fdf-9804-39eb782d6478\") " pod="openshift-machine-config-operator/machine-config-server-ccgf6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668099 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d90a8f89-0b24-47b4-bba6-2b84b7443f78-cert\") pod \"ingress-canary-h695r\" (UID: \"d90a8f89-0b24-47b4-bba6-2b84b7443f78\") " pod="openshift-ingress-canary/ingress-canary-h695r" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668122 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d212b310-5c16-4fdf-9804-39eb782d6478-node-bootstrap-token\") pod \"machine-config-server-ccgf6\" (UID: \"d212b310-5c16-4fdf-9804-39eb782d6478\") " pod="openshift-machine-config-operator/machine-config-server-ccgf6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668140 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4692355-bdb5-4904-86f9-7b01f41aded7-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668160 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd2wc\" (UniqueName: \"kubernetes.io/projected/d90a8f89-0b24-47b4-bba6-2b84b7443f78-kube-api-access-bd2wc\") pod \"ingress-canary-h695r\" (UID: \"d90a8f89-0b24-47b4-bba6-2b84b7443f78\") " pod="openshift-ingress-canary/ingress-canary-h695r" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668177 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4692355-bdb5-4904-86f9-7b01f41aded7-config\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668199 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/93887080-7c19-4f58-bc8d-9eff7323a02e-metrics-tls\") pod \"dns-default-d6n9s\" (UID: \"93887080-7c19-4f58-bc8d-9eff7323a02e\") " pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668219 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6sxw\" (UniqueName: \"kubernetes.io/projected/10ca57c1-d445-4335-8051-f90489bdcacf-kube-api-access-b6sxw\") pod \"kube-storage-version-migrator-operator-b67b599dd-hmnt5\" (UID: \"10ca57c1-d445-4335-8051-f90489bdcacf\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668267 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2rmf\" (UniqueName: \"kubernetes.io/projected/f4692355-bdb5-4904-86f9-7b01f41aded7-kube-api-access-f2rmf\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668287 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgzps\" (UniqueName: \"kubernetes.io/projected/9a2c33ab-1bc7-436d-95b7-098dd2e57b40-kube-api-access-zgzps\") pod \"migrator-59844c95c7-796c7\" (UID: \"9a2c33ab-1bc7-436d-95b7-098dd2e57b40\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-796c7" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668304 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93887080-7c19-4f58-bc8d-9eff7323a02e-config-volume\") pod \"dns-default-d6n9s\" (UID: \"93887080-7c19-4f58-bc8d-9eff7323a02e\") " pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.668320 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10ca57c1-d445-4335-8051-f90489bdcacf-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hmnt5\" (UID: \"10ca57c1-d445-4335-8051-f90489bdcacf\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" Feb 18 15:24:13 crc kubenswrapper[4968]: E0218 15:24:13.669336 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:14.169317075 +0000 UTC m=+153.554761937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.669588 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4692355-bdb5-4904-86f9-7b01f41aded7-config\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.669869 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-csi-data-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.670650 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93887080-7c19-4f58-bc8d-9eff7323a02e-config-volume\") pod \"dns-default-d6n9s\" (UID: \"93887080-7c19-4f58-bc8d-9eff7323a02e\") " pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.670977 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-plugins-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.671057 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-registration-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.671562 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10ca57c1-d445-4335-8051-f90489bdcacf-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hmnt5\" (UID: \"10ca57c1-d445-4335-8051-f90489bdcacf\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.671652 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-socket-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.671958 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c858a456-78e0-4f36-9a4e-6db413d94d48-mountpoint-dir\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.672574 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4692355-bdb5-4904-86f9-7b01f41aded7-service-ca-bundle\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.676660 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/93887080-7c19-4f58-bc8d-9eff7323a02e-metrics-tls\") pod \"dns-default-d6n9s\" (UID: \"93887080-7c19-4f58-bc8d-9eff7323a02e\") " pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.677461 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4692355-bdb5-4904-86f9-7b01f41aded7-serving-cert\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.679109 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d212b310-5c16-4fdf-9804-39eb782d6478-node-bootstrap-token\") pod \"machine-config-server-ccgf6\" (UID: \"d212b310-5c16-4fdf-9804-39eb782d6478\") " pod="openshift-machine-config-operator/machine-config-server-ccgf6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.680735 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d212b310-5c16-4fdf-9804-39eb782d6478-certs\") pod \"machine-config-server-ccgf6\" (UID: \"d212b310-5c16-4fdf-9804-39eb782d6478\") " pod="openshift-machine-config-operator/machine-config-server-ccgf6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.682264 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4692355-bdb5-4904-86f9-7b01f41aded7-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.689045 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10ca57c1-d445-4335-8051-f90489bdcacf-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hmnt5\" (UID: \"10ca57c1-d445-4335-8051-f90489bdcacf\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.689050 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d90a8f89-0b24-47b4-bba6-2b84b7443f78-cert\") pod \"ingress-canary-h695r\" (UID: \"d90a8f89-0b24-47b4-bba6-2b84b7443f78\") " pod="openshift-ingress-canary/ingress-canary-h695r" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.689656 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwqbc\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-kube-api-access-kwqbc\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.705890 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pd85\" (UniqueName: \"kubernetes.io/projected/f1b7574f-d70f-46dd-ac82-2cc55659bf7d-kube-api-access-2pd85\") pod \"ingress-operator-5b745b69d9-pqd7g\" (UID: \"f1b7574f-d70f-46dd-ac82-2cc55659bf7d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.715108 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.722902 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-bound-sa-token\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.741719 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66"] Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.747982 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac98f4f8-45d4-4bce-b635-8be96ba93afd-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-4sklv\" (UID: \"ac98f4f8-45d4-4bce-b635-8be96ba93afd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.768151 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kv8lj\" (UID: \"2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.769315 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:13 crc kubenswrapper[4968]: E0218 15:24:13.769820 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:14.269789844 +0000 UTC m=+153.655234706 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.779895 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwjjb\" (UniqueName: \"kubernetes.io/projected/fb43c877-a415-4362-a629-8b77f5bce23f-kube-api-access-kwjjb\") pod \"service-ca-operator-777779d784-857s6\" (UID: \"fb43c877-a415-4362-a629-8b77f5bce23f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.834858 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6v575"] Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.853005 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk"] Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.852761 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.849159 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.867043 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqqrd\" (UniqueName: \"kubernetes.io/projected/93887080-7c19-4f58-bc8d-9eff7323a02e-kube-api-access-xqqrd\") pod \"dns-default-d6n9s\" (UID: \"93887080-7c19-4f58-bc8d-9eff7323a02e\") " pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.871411 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:13 crc kubenswrapper[4968]: E0218 15:24:13.871917 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:14.371899441 +0000 UTC m=+153.757344303 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.877679 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf885\" (UniqueName: \"kubernetes.io/projected/c858a456-78e0-4f36-9a4e-6db413d94d48-kube-api-access-mf885\") pod \"csi-hostpathplugin-r6dwq\" (UID: \"c858a456-78e0-4f36-9a4e-6db413d94d48\") " pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.878886 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv822\" (UniqueName: \"kubernetes.io/projected/d212b310-5c16-4fdf-9804-39eb782d6478-kube-api-access-wv822\") pod \"machine-config-server-ccgf6\" (UID: \"d212b310-5c16-4fdf-9804-39eb782d6478\") " pod="openshift-machine-config-operator/machine-config-server-ccgf6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.879014 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2rmf\" (UniqueName: \"kubernetes.io/projected/f4692355-bdb5-4904-86f9-7b01f41aded7-kube-api-access-f2rmf\") pod \"authentication-operator-69f744f599-lmz7d\" (UID: \"f4692355-bdb5-4904-86f9-7b01f41aded7\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.890193 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgzps\" (UniqueName: \"kubernetes.io/projected/9a2c33ab-1bc7-436d-95b7-098dd2e57b40-kube-api-access-zgzps\") pod \"migrator-59844c95c7-796c7\" (UID: \"9a2c33ab-1bc7-436d-95b7-098dd2e57b40\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-796c7" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.896862 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.922050 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.922320 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd2wc\" (UniqueName: \"kubernetes.io/projected/d90a8f89-0b24-47b4-bba6-2b84b7443f78-kube-api-access-bd2wc\") pod \"ingress-canary-h695r\" (UID: \"d90a8f89-0b24-47b4-bba6-2b84b7443f78\") " pod="openshift-ingress-canary/ingress-canary-h695r" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.932253 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6sxw\" (UniqueName: \"kubernetes.io/projected/10ca57c1-d445-4335-8051-f90489bdcacf-kube-api-access-b6sxw\") pod \"kube-storage-version-migrator-operator-b67b599dd-hmnt5\" (UID: \"10ca57c1-d445-4335-8051-f90489bdcacf\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" Feb 18 15:24:13 crc kubenswrapper[4968]: W0218 15:24:13.944936 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5814e0fe_15f6_4675_ab06_ca9f9a467cd0.slice/crio-1b8c919fbea395d9e30566cbbec4e23323be51ec0cd159a94619e0df2af3b193 WatchSource:0}: Error finding container 1b8c919fbea395d9e30566cbbec4e23323be51ec0cd159a94619e0df2af3b193: Status 404 returned error can't find the container with id 1b8c919fbea395d9e30566cbbec4e23323be51ec0cd159a94619e0df2af3b193 Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.954426 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.970118 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-796c7" Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.972579 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:13 crc kubenswrapper[4968]: E0218 15:24:13.973297 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:14.473265705 +0000 UTC m=+153.858710567 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:13 crc kubenswrapper[4968]: I0218 15:24:13.980375 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.008891 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.058587 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-h695r" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.068089 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.074491 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.074767 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ccgf6" Feb 18 15:24:14 crc kubenswrapper[4968]: E0218 15:24:14.074967 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:14.57495141 +0000 UTC m=+153.960396272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.095407 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.123933 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-x829v"] Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.137449 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" event={"ID":"29f23275-ee4d-41d6-acaa-a9c1363d71c3","Type":"ContainerStarted","Data":"e0b110667aa46c5ec04f067c659aa0fb695621782401fe1bda6bc0feb472b53c"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.142583 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" event={"ID":"5814e0fe-15f6-4675-ab06-ca9f9a467cd0","Type":"ContainerStarted","Data":"1b8c919fbea395d9e30566cbbec4e23323be51ec0cd159a94619e0df2af3b193"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.145227 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" event={"ID":"210d3bb5-c13d-49fc-9d5c-5b2ce0847d92","Type":"ContainerStarted","Data":"3d86726d56baee7c4fc1c3ee072c5e29fe9ff414bc072520e6869de15c757425"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.146316 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" event={"ID":"39580550-6421-4eb7-8727-42c26ad5c3f2","Type":"ContainerStarted","Data":"0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.146347 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" event={"ID":"39580550-6421-4eb7-8727-42c26ad5c3f2","Type":"ContainerStarted","Data":"3b56b61e95ee58adbf4461ff17d2a859d0ef633b8ac19776bae8bce931f9d84a"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.146907 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.147842 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" event={"ID":"402543c2-2831-4d3b-87ca-6d284c6d9be1","Type":"ContainerStarted","Data":"161cd7ba8bd36c9ee12f6dcafe092b26c6811af1bf4b3dd7ce8ab9bdc9af3998"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.147868 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" event={"ID":"402543c2-2831-4d3b-87ca-6d284c6d9be1","Type":"ContainerStarted","Data":"087e7c32374f281931f4fc189528d6db34c7bd0110f03e3c9f25237df031e7e6"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.148668 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-nkbhx" event={"ID":"41a0f2d1-405f-4396-a081-b874a35bacf8","Type":"ContainerStarted","Data":"67949ca26fe418110301c7baf39da6fe3c5755f5c82c12b7a61a1faa17e433e4"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.149165 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.150565 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" event={"ID":"f57a4a48-39f0-44f3-bbf8-982a27fd6f14","Type":"ContainerStarted","Data":"da609585189ff1e0fc16cbb7b9ede2bd1a49676ae2b350870372d82ea4a0d491"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.150611 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" event={"ID":"f57a4a48-39f0-44f3-bbf8-982a27fd6f14","Type":"ContainerStarted","Data":"1ee8443ee3a0b43dcf64bd0757eb068c98403c66acaa92812a17974ed4fd8489"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.150773 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.151864 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-w9zh9" event={"ID":"1c808375-740d-4fe5-ae91-64f776b28e8a","Type":"ContainerStarted","Data":"c76f1c4aabf0c4f0125ec7f750cdb2bd8dc1e9d1bde26cc878f8220cba157833"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.151897 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-w9zh9" event={"ID":"1c808375-740d-4fe5-ae91-64f776b28e8a","Type":"ContainerStarted","Data":"b7848a5e8d22212e0d6a736e69875ea9cf12873d1a916e5b2a03f176495d0a92"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.160239 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" event={"ID":"d43c1374-e913-423e-ab78-6e14b1c16bfd","Type":"ContainerStarted","Data":"806b103b40e5080a4ee86c0269c6de6b7794809f6902290e96e3635af439a680"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.160328 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" event={"ID":"d43c1374-e913-423e-ab78-6e14b1c16bfd","Type":"ContainerStarted","Data":"1e0c5ac0c54bd92cc40c0a183b91314a06590b85cdc540e2fc503c53981f8235"} Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.160973 4968 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-7b762 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.161025 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" podUID="d8e923a2-9b5c-49db-8a02-5eff8ca31033" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.176403 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:14 crc kubenswrapper[4968]: E0218 15:24:14.176643 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:14.676601132 +0000 UTC m=+154.062045994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.177100 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:14 crc kubenswrapper[4968]: E0218 15:24:14.177579 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:14.67756803 +0000 UTC m=+154.063012892 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.233849 4968 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7kwbt container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" start-of-body= Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.233864 4968 patch_prober.go:28] interesting pod/console-operator-58897d9998-nkbhx container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.233920 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" podUID="f57a4a48-39f0-44f3-bbf8-982a27fd6f14" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.234021 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-nkbhx" podUID="41a0f2d1-405f-4396-a081-b874a35bacf8" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.234383 4968 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-f7s66 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.234879 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" podUID="39580550-6421-4eb7-8727-42c26ad5c3f2" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.278345 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:14 crc kubenswrapper[4968]: E0218 15:24:14.279814 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:14.7797894 +0000 UTC m=+154.165234262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:14 crc kubenswrapper[4968]: W0218 15:24:14.281287 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod954a7c44_9845_4a84_8cd6_a5fe9be2d7d2.slice/crio-23c2d96106c539ae4d8a4dd90d6840212e83953a6d51f37839cb48499143bb8f WatchSource:0}: Error finding container 23c2d96106c539ae4d8a4dd90d6840212e83953a6d51f37839cb48499143bb8f: Status 404 returned error can't find the container with id 23c2d96106c539ae4d8a4dd90d6840212e83953a6d51f37839cb48499143bb8f Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.288002 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.353127 4968 patch_prober.go:28] interesting pod/router-default-5444994796-w9zh9 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.353183 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w9zh9" podUID="1c808375-740d-4fe5-ae91-64f776b28e8a" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.368926 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.368985 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.384435 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:14 crc kubenswrapper[4968]: E0218 15:24:14.384855 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:14.884839693 +0000 UTC m=+154.270284555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.486539 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:14 crc kubenswrapper[4968]: E0218 15:24:14.487550 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:14.987523776 +0000 UTC m=+154.372968638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.561254 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6"] Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.577083 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" podStartSLOduration=133.577056234 podStartE2EDuration="2m13.577056234s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:14.567979088 +0000 UTC m=+153.953423970" watchObservedRunningTime="2026-02-18 15:24:14.577056234 +0000 UTC m=+153.962501096" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.586561 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-whmt5"] Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.588617 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4"] Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.590725 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:14 crc kubenswrapper[4968]: E0218 15:24:14.591198 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:15.091178038 +0000 UTC m=+154.476622900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.594299 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nz8lf"] Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.602098 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6"] Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.610491 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq"] Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.691905 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:14 crc kubenswrapper[4968]: E0218 15:24:14.692483 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:15.192455611 +0000 UTC m=+154.577900473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.723808 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8g9pk"] Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.728039 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-td7d6"] Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.793976 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:14 crc kubenswrapper[4968]: E0218 15:24:14.794683 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:15.29465804 +0000 UTC m=+154.680102902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.799783 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mjtgm" podStartSLOduration=133.799618066 podStartE2EDuration="2m13.799618066s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:14.797867614 +0000 UTC m=+154.183312496" watchObservedRunningTime="2026-02-18 15:24:14.799618066 +0000 UTC m=+154.185062928" Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.895348 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:14 crc kubenswrapper[4968]: E0218 15:24:14.895548 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:15.39550529 +0000 UTC m=+154.780950152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.895733 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:14 crc kubenswrapper[4968]: E0218 15:24:14.896127 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:15.396118448 +0000 UTC m=+154.781563310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:14 crc kubenswrapper[4968]: I0218 15:24:14.997516 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:15 crc kubenswrapper[4968]: E0218 15:24:14.998001 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:15.497970147 +0000 UTC m=+154.883415009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.081264 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" podStartSLOduration=133.0812275 podStartE2EDuration="2m13.0812275s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:15.076370337 +0000 UTC m=+154.461815209" watchObservedRunningTime="2026-02-18 15:24:15.0812275 +0000 UTC m=+154.466672372" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.105449 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:15 crc kubenswrapper[4968]: E0218 15:24:15.105996 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:15.605968336 +0000 UTC m=+154.991413198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.138075 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-nkbhx" podStartSLOduration=134.138048668 podStartE2EDuration="2m14.138048668s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:15.133240596 +0000 UTC m=+154.518685458" watchObservedRunningTime="2026-02-18 15:24:15.138048668 +0000 UTC m=+154.523493520" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.212599 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:15 crc kubenswrapper[4968]: E0218 15:24:15.214254 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:15.714228323 +0000 UTC m=+155.099673185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.217624 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-s6qsc"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.225082 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.260790 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dxl4k"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.260839 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" event={"ID":"954a7c44-9845-4a84-8cd6-a5fe9be2d7d2","Type":"ContainerStarted","Data":"dfcfe3b1a5ff9696899ded987f57ce2097b3b88bea5a328878fff717cb31abfc"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.260863 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" event={"ID":"954a7c44-9845-4a84-8cd6-a5fe9be2d7d2","Type":"ContainerStarted","Data":"23c2d96106c539ae4d8a4dd90d6840212e83953a6d51f37839cb48499143bb8f"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.261861 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" event={"ID":"bdb30229-c0c5-4f40-a69c-6d8548cb76d8","Type":"ContainerStarted","Data":"63b3d5df89ec9e27336700b9c4d9da2041a63683cedff8b370c9ce05626ab9d0"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.263487 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" event={"ID":"210d3bb5-c13d-49fc-9d5c-5b2ce0847d92","Type":"ContainerStarted","Data":"b272d707f2382c5b0b6f378767b600b41798b5576d70e8abe962ec5ad633ee9b"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.266027 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ccgf6" event={"ID":"d212b310-5c16-4fdf-9804-39eb782d6478","Type":"ContainerStarted","Data":"3f7cf785106f182ad41a10632ebf7d1c2e6884ab465bcafb169a0accf0327c83"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.266071 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ccgf6" event={"ID":"d212b310-5c16-4fdf-9804-39eb782d6478","Type":"ContainerStarted","Data":"9df36c2712eef3c835adfd001af58f3eb815d82738e9b8643c98ce0476a18429"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.266817 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" podStartSLOduration=133.266807106 podStartE2EDuration="2m13.266807106s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:15.264728135 +0000 UTC m=+154.650173007" watchObservedRunningTime="2026-02-18 15:24:15.266807106 +0000 UTC m=+154.652251968" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.284894 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" event={"ID":"402543c2-2831-4d3b-87ca-6d284c6d9be1","Type":"ContainerStarted","Data":"318f815255e9992ee55441c9f58330060517e44db11fb10f17eadf497e7909d1"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.287840 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" event={"ID":"29f23275-ee4d-41d6-acaa-a9c1363d71c3","Type":"ContainerStarted","Data":"c4ec87478f52c15d5271261628cc9b6cbbbacf65aed9be29bc611fc0d01e7593"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.288433 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.304728 4968 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-p58wk container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.304836 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" podUID="29f23275-ee4d-41d6-acaa-a9c1363d71c3" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.309217 4968 patch_prober.go:28] interesting pod/router-default-5444994796-w9zh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 15:24:15 crc kubenswrapper[4968]: [-]has-synced failed: reason withheld Feb 18 15:24:15 crc kubenswrapper[4968]: [+]process-running ok Feb 18 15:24:15 crc kubenswrapper[4968]: healthz check failed Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.309288 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w9zh9" podUID="1c808375-740d-4fe5-ae91-64f776b28e8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.316324 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:15 crc kubenswrapper[4968]: E0218 15:24:15.318949 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:15.818929796 +0000 UTC m=+155.204374658 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.324916 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.336844 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.357053 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" event={"ID":"df2ba163-1935-4600-a6c8-dd73d151ed32","Type":"ContainerStarted","Data":"aaf16a5906fa277ca0f32a099e908dd616450afb292fd176f5eb5107a194957c"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.370343 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" event={"ID":"4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121","Type":"ContainerStarted","Data":"64ef13c90cf2d84c91df75aeae0875d0b25cb9be86abbcd19e8a89f5cc16346b"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.417633 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:15 crc kubenswrapper[4968]: E0218 15:24:15.419069 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:15.919029344 +0000 UTC m=+155.304474206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.459377 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.471313 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dnmb8"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.497163 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.501353 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" event={"ID":"5814e0fe-15f6-4675-ab06-ca9f9a467cd0","Type":"ContainerStarted","Data":"c66eb09d26a9e433b4560344e6b2409f87b6d15d867ae6648dbb037bac4df31b"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.514782 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-857s6"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.565111 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:15 crc kubenswrapper[4968]: E0218 15:24:15.567144 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:16.06712205 +0000 UTC m=+155.452566912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.570413 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.579423 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" podStartSLOduration=133.57939026 podStartE2EDuration="2m13.57939026s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:15.447116518 +0000 UTC m=+154.832561390" watchObservedRunningTime="2026-02-18 15:24:15.57939026 +0000 UTC m=+154.964835122" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.582904 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" event={"ID":"57795f27-c9c3-4f38-a5da-364e2488f343","Type":"ContainerStarted","Data":"231f552c329f92d7e13f1ce9560950b542e0b97058f4b23d2263e8f4427e6ffd"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.605126 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-w9zh9" podStartSLOduration=134.605100104 podStartE2EDuration="2m14.605100104s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:15.604297011 +0000 UTC m=+154.989741883" watchObservedRunningTime="2026-02-18 15:24:15.605100104 +0000 UTC m=+154.990544966" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.627202 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-td7d6" event={"ID":"5985084a-4c95-41f7-a40a-50550738e31c","Type":"ContainerStarted","Data":"86ecc85ab52e729c344edd3234b96452e626ee439c995b3a2193a44f7018ee00"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.642658 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-whmt5" event={"ID":"f9194c82-960a-4dea-94ef-5b2d5b4e4a75","Type":"ContainerStarted","Data":"47368a03bc327429dd957c01cbf72284d48f328900dc11172834aba92576447b"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.648767 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nz8lf" event={"ID":"ba1be0b0-ffa7-4b12-bc99-354621c8dd5e","Type":"ContainerStarted","Data":"bfd0dcf13befcfc3a3abb7c43f6f33fa63248ff898b4a2293d0f92206eab8f00"} Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.663822 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-m6p2k" podStartSLOduration=134.663802057 podStartE2EDuration="2m14.663802057s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:15.66221117 +0000 UTC m=+155.047656032" watchObservedRunningTime="2026-02-18 15:24:15.663802057 +0000 UTC m=+155.049246919" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.666984 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-nkbhx" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.667143 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:15 crc kubenswrapper[4968]: E0218 15:24:15.667806 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:16.167775393 +0000 UTC m=+155.553220255 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.674492 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.691801 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ckl6x" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.719784 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" podStartSLOduration=134.719763309 podStartE2EDuration="2m14.719763309s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:15.717276196 +0000 UTC m=+155.102721048" watchObservedRunningTime="2026-02-18 15:24:15.719763309 +0000 UTC m=+155.105208171" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.726112 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.728575 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.738644 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.769050 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.771000 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-lmz7d"] Feb 18 15:24:15 crc kubenswrapper[4968]: E0218 15:24:15.781911 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:16.281887423 +0000 UTC m=+155.667332285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.786959 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-r6dwq"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.792551 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hjsr5" podStartSLOduration=135.792528025 podStartE2EDuration="2m15.792528025s" podCreationTimestamp="2026-02-18 15:22:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:15.786177238 +0000 UTC m=+155.171622110" watchObservedRunningTime="2026-02-18 15:24:15.792528025 +0000 UTC m=+155.177972887" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.793392 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-796c7"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.871462 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:15 crc kubenswrapper[4968]: E0218 15:24:15.872116 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:16.37208384 +0000 UTC m=+155.757528702 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.892067 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7kwbt" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.949836 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" podStartSLOduration=133.949802861 podStartE2EDuration="2m13.949802861s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:15.926921519 +0000 UTC m=+155.312366381" watchObservedRunningTime="2026-02-18 15:24:15.949802861 +0000 UTC m=+155.335247723" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.954247 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-d6n9s"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.976130 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-h695r"] Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.986948 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:15 crc kubenswrapper[4968]: I0218 15:24:15.987331 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4q2wd" podStartSLOduration=134.987304831 podStartE2EDuration="2m14.987304831s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:15.986678043 +0000 UTC m=+155.372122925" watchObservedRunningTime="2026-02-18 15:24:15.987304831 +0000 UTC m=+155.372749693" Feb 18 15:24:15 crc kubenswrapper[4968]: E0218 15:24:15.987410 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:16.487386784 +0000 UTC m=+155.872831646 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.070349 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-h58tr" podStartSLOduration=134.070320548 podStartE2EDuration="2m14.070320548s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:16.061579681 +0000 UTC m=+155.447024543" watchObservedRunningTime="2026-02-18 15:24:16.070320548 +0000 UTC m=+155.455765410" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.073920 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-6v575" podStartSLOduration=135.073901503 podStartE2EDuration="2m15.073901503s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:16.034252499 +0000 UTC m=+155.419697381" watchObservedRunningTime="2026-02-18 15:24:16.073901503 +0000 UTC m=+155.459346365" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.096273 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.099253 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-ccgf6" podStartSLOduration=6.099222506 podStartE2EDuration="6.099222506s" podCreationTimestamp="2026-02-18 15:24:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:16.09562136 +0000 UTC m=+155.481066232" watchObservedRunningTime="2026-02-18 15:24:16.099222506 +0000 UTC m=+155.484667368" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.099408 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:16 crc kubenswrapper[4968]: E0218 15:24:16.100740 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:16.60071996 +0000 UTC m=+155.986164822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.103113 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.103591 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:16 crc kubenswrapper[4968]: E0218 15:24:16.113447 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:16.604854191 +0000 UTC m=+155.990299053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.173439 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.208963 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:16 crc kubenswrapper[4968]: E0218 15:24:16.210645 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:16.710588544 +0000 UTC m=+156.096033406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.304050 4968 patch_prober.go:28] interesting pod/router-default-5444994796-w9zh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 15:24:16 crc kubenswrapper[4968]: [-]has-synced failed: reason withheld Feb 18 15:24:16 crc kubenswrapper[4968]: [+]process-running ok Feb 18 15:24:16 crc kubenswrapper[4968]: healthz check failed Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.304137 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w9zh9" podUID="1c808375-740d-4fe5-ae91-64f776b28e8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.310840 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:16 crc kubenswrapper[4968]: E0218 15:24:16.311346 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:16.811328761 +0000 UTC m=+156.196773623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.416486 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:16 crc kubenswrapper[4968]: E0218 15:24:16.416912 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:16.916887419 +0000 UTC m=+156.302332281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.519403 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:16 crc kubenswrapper[4968]: E0218 15:24:16.520461 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:17.020441398 +0000 UTC m=+156.405886260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.621690 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:16 crc kubenswrapper[4968]: E0218 15:24:16.622152 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:17.122129872 +0000 UTC m=+156.507574734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.722412 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" event={"ID":"bdb30229-c0c5-4f40-a69c-6d8548cb76d8","Type":"ContainerStarted","Data":"e2769c822845fc44891ac3b90674d91a7b464159321b05c47077d11c0fb044ae"} Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.723459 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:16 crc kubenswrapper[4968]: E0218 15:24:16.723876 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:17.223858508 +0000 UTC m=+156.609303370 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.724407 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.727951 4968 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-cx7j6 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.728035 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" podUID="bdb30229-c0c5-4f40-a69c-6d8548cb76d8" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.741591 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" podStartSLOduration=134.741571497 podStartE2EDuration="2m14.741571497s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:16.740997041 +0000 UTC m=+156.126441903" watchObservedRunningTime="2026-02-18 15:24:16.741571497 +0000 UTC m=+156.127016359" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.743634 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" event={"ID":"dfffde2b-573c-418b-903b-9117e958dabc","Type":"ContainerStarted","Data":"907098552f11a00ae57716a0336298618f51520d4d1dd82acc0bb5cc9dd00980"} Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.751499 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-796c7" event={"ID":"9a2c33ab-1bc7-436d-95b7-098dd2e57b40","Type":"ContainerStarted","Data":"edfa1d1f6d1832c46e88ac02e745bff28421c932983974f0f9c00bfcf7ffaf81"} Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.766095 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" event={"ID":"f4692355-bdb5-4904-86f9-7b01f41aded7","Type":"ContainerStarted","Data":"0341f5af64956c2451e1ced26726979dabd59e421d636da2deb95e888433efc3"} Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.805397 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" event={"ID":"f1b7574f-d70f-46dd-ac82-2cc55659bf7d","Type":"ContainerStarted","Data":"4c7c5fb1fe8880e39f8069b3d3b5dd733d5ed0493dca8e531997117e1ddcb1ee"} Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.825528 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:16 crc kubenswrapper[4968]: E0218 15:24:16.828310 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:17.328279262 +0000 UTC m=+156.713724124 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.863152 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-td7d6" event={"ID":"5985084a-4c95-41f7-a40a-50550738e31c","Type":"ContainerStarted","Data":"c69ec63a777d6e3c53bde661d36ca0a92b7571cda41ea6761a129e7ff344ccdc"} Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.864483 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-td7d6" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.871218 4968 patch_prober.go:28] interesting pod/downloads-7954f5f757-td7d6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.871285 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-td7d6" podUID="5985084a-4c95-41f7-a40a-50550738e31c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.880079 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" event={"ID":"10ca57c1-d445-4335-8051-f90489bdcacf","Type":"ContainerStarted","Data":"482743115b7f29f2eeb5efa60c39f7754b784ca49c0025230ef3e5eb88c789d3"} Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.896621 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-td7d6" podStartSLOduration=135.896596487 podStartE2EDuration="2m15.896596487s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:16.894505845 +0000 UTC m=+156.279950707" watchObservedRunningTime="2026-02-18 15:24:16.896596487 +0000 UTC m=+156.282041349" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.908993 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" event={"ID":"57795f27-c9c3-4f38-a5da-364e2488f343","Type":"ContainerStarted","Data":"9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5"} Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.909472 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.911098 4968 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8g9pk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.38:6443/healthz\": dial tcp 10.217.0.38:6443: connect: connection refused" start-of-body= Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.911144 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" podUID="57795f27-c9c3-4f38-a5da-364e2488f343" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.38:6443/healthz\": dial tcp 10.217.0.38:6443: connect: connection refused" Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.926946 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:16 crc kubenswrapper[4968]: E0218 15:24:16.927390 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:17.42737523 +0000 UTC m=+156.812820092 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.938383 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d6n9s" event={"ID":"93887080-7c19-4f58-bc8d-9eff7323a02e","Type":"ContainerStarted","Data":"f7b53f9c3de5d28aba81f1c103bd608833b25cc90b4dd4ac4275dbf4472d1d5c"} Feb 18 15:24:16 crc kubenswrapper[4968]: I0218 15:24:16.947023 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" podStartSLOduration=135.946989276 podStartE2EDuration="2m15.946989276s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:16.941955208 +0000 UTC m=+156.327400070" watchObservedRunningTime="2026-02-18 15:24:16.946989276 +0000 UTC m=+156.332434138" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.030551 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:17 crc kubenswrapper[4968]: E0218 15:24:17.032124 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:17.532099263 +0000 UTC m=+156.917544125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.044953 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" event={"ID":"4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121","Type":"ContainerStarted","Data":"8538d16f3253e775c56c1e9d4166d509cc648386deffc4ac31d266924f680f53"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.045914 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.072539 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" event={"ID":"954a7c44-9845-4a84-8cd6-a5fe9be2d7d2","Type":"ContainerStarted","Data":"f990d027d229ee7ef6339b49b51df1f9ec00e7421eca07dbbeb9fc35354639e6"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.078054 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" podStartSLOduration=135.078030491 podStartE2EDuration="2m15.078030491s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:17.073763506 +0000 UTC m=+156.459208378" watchObservedRunningTime="2026-02-18 15:24:17.078030491 +0000 UTC m=+156.463475353" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.115025 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-x829v" podStartSLOduration=135.114988146 podStartE2EDuration="2m15.114988146s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:17.110547146 +0000 UTC m=+156.495992008" watchObservedRunningTime="2026-02-18 15:24:17.114988146 +0000 UTC m=+156.500433008" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.140154 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6" event={"ID":"6c8c2a1f-4652-4791-a114-ebb9a46659ec","Type":"ContainerStarted","Data":"8536c65d69ba930645e88e84fd59ec2034f3364ea7b9de2798619df7e5e4fe45"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.140220 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6" event={"ID":"6c8c2a1f-4652-4791-a114-ebb9a46659ec","Type":"ContainerStarted","Data":"4fc777efa2de42d9f240c70c7856f33594003d4655069ac6640937bae8ce65dc"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.141350 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:17 crc kubenswrapper[4968]: E0218 15:24:17.148932 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:17.648910662 +0000 UTC m=+157.034355524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.169399 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" event={"ID":"e058a32b-9d0b-449a-a925-dc19858e6546","Type":"ContainerStarted","Data":"ac7d5129ebc140491f7ee2975ec60e24de53e461ea9868b2fd637091f740a614"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.169462 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" event={"ID":"e058a32b-9d0b-449a-a925-dc19858e6546","Type":"ContainerStarted","Data":"2e7936df6f9ef6b1c8be4cec0daef2ef8284cc0e61fcd0bd99885480b5532f37"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.178440 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" event={"ID":"ac98f4f8-45d4-4bce-b635-8be96ba93afd","Type":"ContainerStarted","Data":"2a89faa5719d5e38bf475cdf526686192b76d2c5d920ab7fff3b72244d4e3876"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.190285 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s6qsc" event={"ID":"2c7da228-335d-4f50-83e6-4ae152ed4874","Type":"ContainerStarted","Data":"d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.190331 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s6qsc" event={"ID":"2c7da228-335d-4f50-83e6-4ae152ed4874","Type":"ContainerStarted","Data":"000bac82fa9da877c36eb7246df400c7db6a60a1846dfbb811a8bfc0cc4516b0"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.194983 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bdhng" podStartSLOduration=137.194944373 podStartE2EDuration="2m17.194944373s" podCreationTimestamp="2026-02-18 15:22:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:17.19007725 +0000 UTC m=+156.575522112" watchObservedRunningTime="2026-02-18 15:24:17.194944373 +0000 UTC m=+156.580389255" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.217582 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" event={"ID":"cc03d1b2-f3f3-4b18-b108-e1a166a6a250","Type":"ContainerStarted","Data":"e58e337ace8777153cb1a3407c99e2c459678535ffd1e4d58842bd61bb53440d"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.218572 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.228965 4968 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dnmb8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/healthz\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.229036 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" podUID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.21:8080/healthz\": dial tcp 10.217.0.21:8080: connect: connection refused" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.249432 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:17 crc kubenswrapper[4968]: E0218 15:24:17.250522 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:17.750499773 +0000 UTC m=+157.135944635 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.258827 4968 generic.go:334] "Generic (PLEG): container finished" podID="f9194c82-960a-4dea-94ef-5b2d5b4e4a75" containerID="8c366ca358b8fb3428f680eb9b331e8b03abeb372843e4a721228bf7c995a9b6" exitCode=0 Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.268258 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-whmt5" event={"ID":"f9194c82-960a-4dea-94ef-5b2d5b4e4a75","Type":"ContainerDied","Data":"8c366ca358b8fb3428f680eb9b331e8b03abeb372843e4a721228bf7c995a9b6"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.287170 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-s6qsc" podStartSLOduration=136.287134519 podStartE2EDuration="2m16.287134519s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:17.233440563 +0000 UTC m=+156.618885425" watchObservedRunningTime="2026-02-18 15:24:17.287134519 +0000 UTC m=+156.672579381" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.315186 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" event={"ID":"df2ba163-1935-4600-a6c8-dd73d151ed32","Type":"ContainerStarted","Data":"24a76a6453364bb533f9290a99bed758d12837dc977c57ac28538bc1fbbad3b4"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.359000 4968 patch_prober.go:28] interesting pod/router-default-5444994796-w9zh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 15:24:17 crc kubenswrapper[4968]: [-]has-synced failed: reason withheld Feb 18 15:24:17 crc kubenswrapper[4968]: [+]process-running ok Feb 18 15:24:17 crc kubenswrapper[4968]: healthz check failed Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.359241 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w9zh9" podUID="1c808375-740d-4fe5-ae91-64f776b28e8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.373688 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:17 crc kubenswrapper[4968]: E0218 15:24:17.376571 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:17.876552073 +0000 UTC m=+157.261996925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.378172 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" podStartSLOduration=135.37815284 podStartE2EDuration="2m15.37815284s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:17.311019949 +0000 UTC m=+156.696464801" watchObservedRunningTime="2026-02-18 15:24:17.37815284 +0000 UTC m=+156.763597702" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.398423 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" event={"ID":"46d0135b-0668-4024-aaee-8f28eb43f235","Type":"ContainerStarted","Data":"76fb93906b456e5a0442b5d0b9bad8a8e59a8b88b2c4f75c58a864c51b7bd98b"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.402741 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" podStartSLOduration=136.402716681 podStartE2EDuration="2m16.402716681s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:17.38702353 +0000 UTC m=+156.772468392" watchObservedRunningTime="2026-02-18 15:24:17.402716681 +0000 UTC m=+156.788161543" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.435758 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" event={"ID":"c858a456-78e0-4f36-9a4e-6db413d94d48","Type":"ContainerStarted","Data":"7f0a665ff79b3a6f9d434885cb244052d1e521dcdd3016928ea403b19c748701"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.450262 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt" event={"ID":"605d503b-946e-47e6-b08f-c6eaf0c3eb9a","Type":"ContainerStarted","Data":"6a0e37df7a6db92e1afd7a27a5969701bfa4ec2c177a846c9d7448c757db8998"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.476449 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:17 crc kubenswrapper[4968]: E0218 15:24:17.476831 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:17.976789034 +0000 UTC m=+157.362233896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.490783 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt" podStartSLOduration=135.490736034 podStartE2EDuration="2m15.490736034s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:17.482021848 +0000 UTC m=+156.867466710" watchObservedRunningTime="2026-02-18 15:24:17.490736034 +0000 UTC m=+156.876180906" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.510819 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nz8lf" event={"ID":"ba1be0b0-ffa7-4b12-bc99-354621c8dd5e","Type":"ContainerStarted","Data":"7587ae556fa54e20c01ccbc4db61bef1039645ededbba4eccbfb53ac2f6b862e"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.554431 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" event={"ID":"e122baf6-f5c4-4f82-87ee-520e14095f73","Type":"ContainerStarted","Data":"edfee93c4e0d3466874101141ef9db00b7f23fdb6ffbba0db2c1d1dda05ae71c"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.569158 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-h695r" event={"ID":"d90a8f89-0b24-47b4-bba6-2b84b7443f78","Type":"ContainerStarted","Data":"cd35ea00e90abc7f00ea2dcbc7b84a187810d59cc9e0401229909e0681cb2b73"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.576251 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" event={"ID":"2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5","Type":"ContainerStarted","Data":"b40ec15ceeec06036444232ea69bf8e6d547427293e958eaef3b75568587fac8"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.578738 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:17 crc kubenswrapper[4968]: E0218 15:24:17.580587 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:18.08056722 +0000 UTC m=+157.466012072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.612008 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" event={"ID":"fb43c877-a415-4362-a629-8b77f5bce23f","Type":"ContainerStarted","Data":"4a27a7b128869d67cc2f4d2b911f5fda950e9cc4568400d7fa3e4b41968067d8"} Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.632383 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qplwp" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.637554 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" podStartSLOduration=135.637528732 podStartE2EDuration="2m15.637528732s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:17.592728767 +0000 UTC m=+156.978173639" watchObservedRunningTime="2026-02-18 15:24:17.637528732 +0000 UTC m=+157.022973594" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.638121 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p58wk" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.639485 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-h695r" podStartSLOduration=7.639470439 podStartE2EDuration="7.639470439s" podCreationTimestamp="2026-02-18 15:24:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:17.638440938 +0000 UTC m=+157.023885800" watchObservedRunningTime="2026-02-18 15:24:17.639470439 +0000 UTC m=+157.024915311" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.690148 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:17 crc kubenswrapper[4968]: E0218 15:24:17.691676 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:18.19164897 +0000 UTC m=+157.577093842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.796385 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.797042 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" podStartSLOduration=135.797020311 podStartE2EDuration="2m15.797020311s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:17.795126166 +0000 UTC m=+157.180571028" watchObservedRunningTime="2026-02-18 15:24:17.797020311 +0000 UTC m=+157.182465173" Feb 18 15:24:17 crc kubenswrapper[4968]: E0218 15:24:17.802511 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:18.302490512 +0000 UTC m=+157.687935374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:17 crc kubenswrapper[4968]: I0218 15:24:17.908856 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:17 crc kubenswrapper[4968]: E0218 15:24:17.909299 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:18.409278336 +0000 UTC m=+157.794723198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.011730 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.012168 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:18.512147305 +0000 UTC m=+157.897592167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.113003 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.113271 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:18.613224001 +0000 UTC m=+157.998668883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.113373 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.113782 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:18.613771977 +0000 UTC m=+157.999216919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.215327 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.215523 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:18.715484092 +0000 UTC m=+158.100928954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.216313 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.216901 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:18.716878063 +0000 UTC m=+158.102322925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.298320 4968 patch_prober.go:28] interesting pod/router-default-5444994796-w9zh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 15:24:18 crc kubenswrapper[4968]: [-]has-synced failed: reason withheld Feb 18 15:24:18 crc kubenswrapper[4968]: [+]process-running ok Feb 18 15:24:18 crc kubenswrapper[4968]: healthz check failed Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.298403 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w9zh9" podUID="1c808375-740d-4fe5-ae91-64f776b28e8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.317932 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.318392 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:18.818369742 +0000 UTC m=+158.203814604 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.420076 4968 csr.go:261] certificate signing request csr-s65fg is approved, waiting to be issued Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.420511 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.421099 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:18.921082126 +0000 UTC m=+158.306526988 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.439357 4968 csr.go:257] certificate signing request csr-s65fg is issued Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.521943 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.522369 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:19.022320087 +0000 UTC m=+158.407764949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.522876 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.523416 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:19.023396819 +0000 UTC m=+158.408841741 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.622664 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nz8lf" event={"ID":"ba1be0b0-ffa7-4b12-bc99-354621c8dd5e","Type":"ContainerStarted","Data":"45c171fcad188802e63c23227206e1bef59bf53362546484e44a27a295ef91df"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.623796 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.624255 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:19.124226938 +0000 UTC m=+158.509671800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.626183 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" event={"ID":"ac98f4f8-45d4-4bce-b635-8be96ba93afd","Type":"ContainerStarted","Data":"442de77e54587fea6e8800fe7ad4c87ae9ae994a62ff8868c715ff634515d059"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.628523 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" event={"ID":"2b8b3fd9-ac3b-4ae5-9c88-9f4cd67a93a5","Type":"ContainerStarted","Data":"6a1530d4b97c6820b8eb7f9b47c9ded0bf1c438a16d55a9a218c3bd8a063a72d"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.646168 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-h695r" event={"ID":"d90a8f89-0b24-47b4-bba6-2b84b7443f78","Type":"ContainerStarted","Data":"36391c4b5be51aeebd1c6854cfe09538adae48a97e17c9713b6ed5cb46b00524"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.659719 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-nz8lf" podStartSLOduration=136.659690618 podStartE2EDuration="2m16.659690618s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:18.655968959 +0000 UTC m=+158.041413831" watchObservedRunningTime="2026-02-18 15:24:18.659690618 +0000 UTC m=+158.045135480" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.660573 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-whmt5" event={"ID":"f9194c82-960a-4dea-94ef-5b2d5b4e4a75","Type":"ContainerStarted","Data":"54ec465df4c984b93f5387133e0be3c28f30f8ed5593625183d1d3da94b94058"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.660673 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-whmt5" event={"ID":"f9194c82-960a-4dea-94ef-5b2d5b4e4a75","Type":"ContainerStarted","Data":"0d05724d3cdf921fd7369092d848b47c869bda6e2700f3f486b321eb98556df3"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.665372 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" event={"ID":"c858a456-78e0-4f36-9a4e-6db413d94d48","Type":"ContainerStarted","Data":"6570733b791147660614e6a01cbdedfe95741bec453c4fff3eed2648b85b81d5"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.671467 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" event={"ID":"dfffde2b-573c-418b-903b-9117e958dabc","Type":"ContainerStarted","Data":"10ef20e1418880b8f91391f76dff29b1a52a283a2c34a898a03b121187155357"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.688218 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" event={"ID":"cc03d1b2-f3f3-4b18-b108-e1a166a6a250","Type":"ContainerStarted","Data":"a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.701151 4968 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dnmb8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/healthz\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.708524 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" podUID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.21:8080/healthz\": dial tcp 10.217.0.21:8080: connect: connection refused" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.711135 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kv8lj" podStartSLOduration=137.711087127 podStartE2EDuration="2m17.711087127s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:18.679647624 +0000 UTC m=+158.065092486" watchObservedRunningTime="2026-02-18 15:24:18.711087127 +0000 UTC m=+158.096531989" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.715532 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-796c7" event={"ID":"9a2c33ab-1bc7-436d-95b7-098dd2e57b40","Type":"ContainerStarted","Data":"1fac8c6ec3fcaf282b439deb9ee2a496a39c602bc02fd00b5f1a26ee7508f940"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.715600 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-796c7" event={"ID":"9a2c33ab-1bc7-436d-95b7-098dd2e57b40","Type":"ContainerStarted","Data":"ee5ec48bcecdf7a40c98782d4bf4255b3185e2d531db4698d1bda931e66cfe6e"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.720117 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" event={"ID":"f4692355-bdb5-4904-86f9-7b01f41aded7","Type":"ContainerStarted","Data":"9b13bcbd50d3a11f1edc25f2f91fdad1b1113775268a3b3201a8777cdbfc60de"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.724538 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-857s6" event={"ID":"fb43c877-a415-4362-a629-8b77f5bce23f","Type":"ContainerStarted","Data":"59cdd6257e9be4fa248b2311a887480c698ee932ca541706926247e677f6da7a"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.725336 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.729795 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:19.229771685 +0000 UTC m=+158.615216547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.752460 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6" event={"ID":"6c8c2a1f-4652-4791-a114-ebb9a46659ec","Type":"ContainerStarted","Data":"e00131c585f5dc66f1d3ef5bdf8345b6ee7a4d8589a233dc5e9dffc3074abd79"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.760317 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4sklv" podStartSLOduration=136.760288161 podStartE2EDuration="2m16.760288161s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:18.720275367 +0000 UTC m=+158.105720229" watchObservedRunningTime="2026-02-18 15:24:18.760288161 +0000 UTC m=+158.145733023" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.763376 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-796c7" podStartSLOduration=137.763363321 podStartE2EDuration="2m17.763363321s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:18.762350281 +0000 UTC m=+158.147795143" watchObservedRunningTime="2026-02-18 15:24:18.763363321 +0000 UTC m=+158.148808183" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.777045 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" event={"ID":"46d0135b-0668-4024-aaee-8f28eb43f235","Type":"ContainerStarted","Data":"099ab1bea54939486b633fd0b9b994021d5c98edb7551ca58c05ad674809f6ff"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.777117 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" event={"ID":"46d0135b-0668-4024-aaee-8f28eb43f235","Type":"ContainerStarted","Data":"fde18e5f965d14d0e36c56bd9cc85d5d5c34fa6ffe3820cb211f5f7e19c54089"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.800516 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-lmz7d" podStartSLOduration=137.800495801 podStartE2EDuration="2m17.800495801s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:18.799480561 +0000 UTC m=+158.184925423" watchObservedRunningTime="2026-02-18 15:24:18.800495801 +0000 UTC m=+158.185940663" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.811720 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x9mtt" event={"ID":"605d503b-946e-47e6-b08f-c6eaf0c3eb9a","Type":"ContainerStarted","Data":"2ccaa8bd66c44ca03215e4635bad3fe24340825b8aa8771151b77289400f74be"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.834500 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.834902 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dxl4k" event={"ID":"e122baf6-f5c4-4f82-87ee-520e14095f73","Type":"ContainerStarted","Data":"2291af9ac63a16397c70a81ba404a19c76dd110b71a6833b63c82ca3f86fe7a0"} Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.849250 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:19.349222411 +0000 UTC m=+158.734667273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.857434 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d6n9s" event={"ID":"93887080-7c19-4f58-bc8d-9eff7323a02e","Type":"ContainerStarted","Data":"5ae662b035100edd020d8e2dba5fba67940cae14d0f2002d435df943b8898944"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.857488 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d6n9s" event={"ID":"93887080-7c19-4f58-bc8d-9eff7323a02e","Type":"ContainerStarted","Data":"1cfb61024bc19db64ef566ee5745b97333d0ad57444b8e1a87505db635b70bfb"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.858223 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.869549 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mncr8" podStartSLOduration=137.869523506 podStartE2EDuration="2m17.869523506s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:18.868227748 +0000 UTC m=+158.253672610" watchObservedRunningTime="2026-02-18 15:24:18.869523506 +0000 UTC m=+158.254968368" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.870628 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-whmt5" podStartSLOduration=137.870619409 podStartE2EDuration="2m17.870619409s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:18.844547693 +0000 UTC m=+158.229992555" watchObservedRunningTime="2026-02-18 15:24:18.870619409 +0000 UTC m=+158.256064271" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.885646 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" event={"ID":"f1b7574f-d70f-46dd-ac82-2cc55659bf7d","Type":"ContainerStarted","Data":"1bc21d0761b50137c1605c8b905daf1840c1d86d65b298314c6bf40256cfa497"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.886174 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" event={"ID":"f1b7574f-d70f-46dd-ac82-2cc55659bf7d","Type":"ContainerStarted","Data":"5b17186d439791679e3464eb38eed957d45900d75652bd475d1e6ab6a8fda4d8"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.890275 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hkb2r"] Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.891400 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.895182 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.904368 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" event={"ID":"4c19a3b5-d1a8-4ac2-bee9-1fbe0187e121","Type":"ContainerStarted","Data":"921feafad44a04541cf3140fc048adc06aa6aa0734eed7224e5d6e301c031622"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.914451 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hkb2r"] Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.919589 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q5lx6" podStartSLOduration=138.919572685 podStartE2EDuration="2m18.919572685s" podCreationTimestamp="2026-02-18 15:22:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:18.904022249 +0000 UTC m=+158.289467121" watchObservedRunningTime="2026-02-18 15:24:18.919572685 +0000 UTC m=+158.305017547" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.934664 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" event={"ID":"10ca57c1-d445-4335-8051-f90489bdcacf","Type":"ContainerStarted","Data":"d214fd83d2ccdb14c5f61d38f7c7cbc4cee527cb9d20a0dd472cffffd56aa731"} Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.937970 4968 patch_prober.go:28] interesting pod/downloads-7954f5f757-td7d6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.938057 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-td7d6" podUID="5985084a-4c95-41f7-a40a-50550738e31c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.940901 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:18 crc kubenswrapper[4968]: E0218 15:24:18.944470 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:19.444446665 +0000 UTC m=+158.829891527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.973047 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.973111 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cx7j6" Feb 18 15:24:18 crc kubenswrapper[4968]: I0218 15:24:18.979569 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k64ds" podStartSLOduration=136.979537485 podStartE2EDuration="2m16.979537485s" podCreationTimestamp="2026-02-18 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:18.950786511 +0000 UTC m=+158.336231373" watchObservedRunningTime="2026-02-18 15:24:18.979537485 +0000 UTC m=+158.364982377" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.004217 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-d6n9s" podStartSLOduration=9.004187778 podStartE2EDuration="9.004187778s" podCreationTimestamp="2026-02-18 15:24:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:18.995534335 +0000 UTC m=+158.380979197" watchObservedRunningTime="2026-02-18 15:24:19.004187778 +0000 UTC m=+158.389632640" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.044979 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:19 crc kubenswrapper[4968]: E0218 15:24:19.045237 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:19.545189442 +0000 UTC m=+158.930634304 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.046235 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83659e7b-5ef2-4128-9a9e-4cbe0e332654-catalog-content\") pod \"community-operators-hkb2r\" (UID: \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\") " pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.046666 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.046957 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83659e7b-5ef2-4128-9a9e-4cbe0e332654-utilities\") pod \"community-operators-hkb2r\" (UID: \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\") " pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.047298 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrrhc\" (UniqueName: \"kubernetes.io/projected/83659e7b-5ef2-4128-9a9e-4cbe0e332654-kube-api-access-lrrhc\") pod \"community-operators-hkb2r\" (UID: \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\") " pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:24:19 crc kubenswrapper[4968]: E0218 15:24:19.049103 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:19.549080526 +0000 UTC m=+158.934525388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.064127 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hmnt5" podStartSLOduration=138.064101057 podStartE2EDuration="2m18.064101057s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:19.055627288 +0000 UTC m=+158.441072150" watchObservedRunningTime="2026-02-18 15:24:19.064101057 +0000 UTC m=+158.449545919" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.093670 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4r24p"] Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.094856 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.104433 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.107167 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4r24p"] Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.152895 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.153172 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83659e7b-5ef2-4128-9a9e-4cbe0e332654-utilities\") pod \"community-operators-hkb2r\" (UID: \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\") " pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.153212 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrrhc\" (UniqueName: \"kubernetes.io/projected/83659e7b-5ef2-4128-9a9e-4cbe0e332654-kube-api-access-lrrhc\") pod \"community-operators-hkb2r\" (UID: \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\") " pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.153260 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83659e7b-5ef2-4128-9a9e-4cbe0e332654-catalog-content\") pod \"community-operators-hkb2r\" (UID: \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\") " pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.153763 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83659e7b-5ef2-4128-9a9e-4cbe0e332654-catalog-content\") pod \"community-operators-hkb2r\" (UID: \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\") " pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:24:19 crc kubenswrapper[4968]: E0218 15:24:19.153860 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:19.65383348 +0000 UTC m=+159.039278332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.154361 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83659e7b-5ef2-4128-9a9e-4cbe0e332654-utilities\") pod \"community-operators-hkb2r\" (UID: \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\") " pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.187594 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pqd7g" podStartSLOduration=138.18756152 podStartE2EDuration="2m18.18756152s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:19.186217431 +0000 UTC m=+158.571662293" watchObservedRunningTime="2026-02-18 15:24:19.18756152 +0000 UTC m=+158.573006382" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.191617 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrrhc\" (UniqueName: \"kubernetes.io/projected/83659e7b-5ef2-4128-9a9e-4cbe0e332654-kube-api-access-lrrhc\") pod \"community-operators-hkb2r\" (UID: \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\") " pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.233762 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.262161 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrpnd\" (UniqueName: \"kubernetes.io/projected/deeab883-52a5-49df-9b71-e3927cf0f2ea-kube-api-access-hrpnd\") pod \"certified-operators-4r24p\" (UID: \"deeab883-52a5-49df-9b71-e3927cf0f2ea\") " pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.279271 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deeab883-52a5-49df-9b71-e3927cf0f2ea-catalog-content\") pod \"certified-operators-4r24p\" (UID: \"deeab883-52a5-49df-9b71-e3927cf0f2ea\") " pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.279319 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.279437 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deeab883-52a5-49df-9b71-e3927cf0f2ea-utilities\") pod \"certified-operators-4r24p\" (UID: \"deeab883-52a5-49df-9b71-e3927cf0f2ea\") " pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:24:19 crc kubenswrapper[4968]: E0218 15:24:19.279962 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:19.779942031 +0000 UTC m=+159.165386893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.325421 4968 patch_prober.go:28] interesting pod/router-default-5444994796-w9zh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 15:24:19 crc kubenswrapper[4968]: [-]has-synced failed: reason withheld Feb 18 15:24:19 crc kubenswrapper[4968]: [+]process-running ok Feb 18 15:24:19 crc kubenswrapper[4968]: healthz check failed Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.325508 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w9zh9" podUID="1c808375-740d-4fe5-ae91-64f776b28e8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.340907 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jvwq9"] Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.342706 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.345628 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jvwq9"] Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.385487 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.385796 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrpnd\" (UniqueName: \"kubernetes.io/projected/deeab883-52a5-49df-9b71-e3927cf0f2ea-kube-api-access-hrpnd\") pod \"certified-operators-4r24p\" (UID: \"deeab883-52a5-49df-9b71-e3927cf0f2ea\") " pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.385831 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/190ec188-7d70-42fd-9bca-1d8d6448d005-catalog-content\") pod \"community-operators-jvwq9\" (UID: \"190ec188-7d70-42fd-9bca-1d8d6448d005\") " pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.385865 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deeab883-52a5-49df-9b71-e3927cf0f2ea-catalog-content\") pod \"certified-operators-4r24p\" (UID: \"deeab883-52a5-49df-9b71-e3927cf0f2ea\") " pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.385913 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/190ec188-7d70-42fd-9bca-1d8d6448d005-utilities\") pod \"community-operators-jvwq9\" (UID: \"190ec188-7d70-42fd-9bca-1d8d6448d005\") " pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.385947 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deeab883-52a5-49df-9b71-e3927cf0f2ea-utilities\") pod \"certified-operators-4r24p\" (UID: \"deeab883-52a5-49df-9b71-e3927cf0f2ea\") " pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.385964 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjhbt\" (UniqueName: \"kubernetes.io/projected/190ec188-7d70-42fd-9bca-1d8d6448d005-kube-api-access-kjhbt\") pod \"community-operators-jvwq9\" (UID: \"190ec188-7d70-42fd-9bca-1d8d6448d005\") " pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:24:19 crc kubenswrapper[4968]: E0218 15:24:19.386097 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:19.886073066 +0000 UTC m=+159.271517928 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.386838 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deeab883-52a5-49df-9b71-e3927cf0f2ea-catalog-content\") pod \"certified-operators-4r24p\" (UID: \"deeab883-52a5-49df-9b71-e3927cf0f2ea\") " pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.387091 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deeab883-52a5-49df-9b71-e3927cf0f2ea-utilities\") pod \"certified-operators-4r24p\" (UID: \"deeab883-52a5-49df-9b71-e3927cf0f2ea\") " pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.442577 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-18 15:19:18 +0000 UTC, rotation deadline is 2026-11-22 01:32:55.527532844 +0000 UTC Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.442634 4968 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6634h8m36.084900638s for next certificate rotation Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.445849 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrpnd\" (UniqueName: \"kubernetes.io/projected/deeab883-52a5-49df-9b71-e3927cf0f2ea-kube-api-access-hrpnd\") pod \"certified-operators-4r24p\" (UID: \"deeab883-52a5-49df-9b71-e3927cf0f2ea\") " pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.488078 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/190ec188-7d70-42fd-9bca-1d8d6448d005-catalog-content\") pod \"community-operators-jvwq9\" (UID: \"190ec188-7d70-42fd-9bca-1d8d6448d005\") " pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.488136 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.488170 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/190ec188-7d70-42fd-9bca-1d8d6448d005-utilities\") pod \"community-operators-jvwq9\" (UID: \"190ec188-7d70-42fd-9bca-1d8d6448d005\") " pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.488206 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjhbt\" (UniqueName: \"kubernetes.io/projected/190ec188-7d70-42fd-9bca-1d8d6448d005-kube-api-access-kjhbt\") pod \"community-operators-jvwq9\" (UID: \"190ec188-7d70-42fd-9bca-1d8d6448d005\") " pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.488983 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/190ec188-7d70-42fd-9bca-1d8d6448d005-catalog-content\") pod \"community-operators-jvwq9\" (UID: \"190ec188-7d70-42fd-9bca-1d8d6448d005\") " pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.489138 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w8dw5"] Feb 18 15:24:19 crc kubenswrapper[4968]: E0218 15:24:19.489286 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:19.989272614 +0000 UTC m=+159.374717476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.489716 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/190ec188-7d70-42fd-9bca-1d8d6448d005-utilities\") pod \"community-operators-jvwq9\" (UID: \"190ec188-7d70-42fd-9bca-1d8d6448d005\") " pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.490351 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.543721 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjhbt\" (UniqueName: \"kubernetes.io/projected/190ec188-7d70-42fd-9bca-1d8d6448d005-kube-api-access-kjhbt\") pod \"community-operators-jvwq9\" (UID: \"190ec188-7d70-42fd-9bca-1d8d6448d005\") " pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.577716 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w8dw5"] Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.589290 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:19 crc kubenswrapper[4968]: E0218 15:24:19.589678 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:20.0896526 +0000 UTC m=+159.475097462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.690631 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.691147 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58srz\" (UniqueName: \"kubernetes.io/projected/e8871835-fc7c-4c43-a1f2-1450cab78bfe-kube-api-access-58srz\") pod \"certified-operators-w8dw5\" (UID: \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\") " pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.691226 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8871835-fc7c-4c43-a1f2-1450cab78bfe-utilities\") pod \"certified-operators-w8dw5\" (UID: \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\") " pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.691274 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8871835-fc7c-4c43-a1f2-1450cab78bfe-catalog-content\") pod \"certified-operators-w8dw5\" (UID: \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\") " pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:24:19 crc kubenswrapper[4968]: E0218 15:24:19.692245 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:20.192228571 +0000 UTC m=+159.577673433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.705584 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.745158 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.795970 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.796072 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8871835-fc7c-4c43-a1f2-1450cab78bfe-utilities\") pod \"certified-operators-w8dw5\" (UID: \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\") " pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.796116 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8871835-fc7c-4c43-a1f2-1450cab78bfe-catalog-content\") pod \"certified-operators-w8dw5\" (UID: \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\") " pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.796183 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58srz\" (UniqueName: \"kubernetes.io/projected/e8871835-fc7c-4c43-a1f2-1450cab78bfe-kube-api-access-58srz\") pod \"certified-operators-w8dw5\" (UID: \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\") " pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:24:19 crc kubenswrapper[4968]: E0218 15:24:19.796653 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:20.296636025 +0000 UTC m=+159.682080887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.797108 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8871835-fc7c-4c43-a1f2-1450cab78bfe-utilities\") pod \"certified-operators-w8dw5\" (UID: \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\") " pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.797363 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8871835-fc7c-4c43-a1f2-1450cab78bfe-catalog-content\") pod \"certified-operators-w8dw5\" (UID: \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\") " pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.834708 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58srz\" (UniqueName: \"kubernetes.io/projected/e8871835-fc7c-4c43-a1f2-1450cab78bfe-kube-api-access-58srz\") pod \"certified-operators-w8dw5\" (UID: \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\") " pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.932950 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:19 crc kubenswrapper[4968]: E0218 15:24:19.933562 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:20.433535292 +0000 UTC m=+159.818980324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.948880 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hkb2r"] Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.978916 4968 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dnmb8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/healthz\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.978992 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" podUID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.21:8080/healthz\": dial tcp 10.217.0.21:8080: connect: connection refused" Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.979187 4968 patch_prober.go:28] interesting pod/downloads-7954f5f757-td7d6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 18 15:24:19 crc kubenswrapper[4968]: I0218 15:24:19.979271 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-td7d6" podUID="5985084a-4c95-41f7-a40a-50550738e31c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.037569 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:20 crc kubenswrapper[4968]: E0218 15:24:20.039428 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:20.539394799 +0000 UTC m=+159.924839711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.111605 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.141797 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:20 crc kubenswrapper[4968]: E0218 15:24:20.142136 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:20.642118664 +0000 UTC m=+160.027563526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.247383 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:20 crc kubenswrapper[4968]: E0218 15:24:20.248553 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:20.748529017 +0000 UTC m=+160.133973879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.298094 4968 patch_prober.go:28] interesting pod/router-default-5444994796-w9zh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 15:24:20 crc kubenswrapper[4968]: [-]has-synced failed: reason withheld Feb 18 15:24:20 crc kubenswrapper[4968]: [+]process-running ok Feb 18 15:24:20 crc kubenswrapper[4968]: healthz check failed Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.298201 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w9zh9" podUID="1c808375-740d-4fe5-ae91-64f776b28e8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.353600 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:20 crc kubenswrapper[4968]: E0218 15:24:20.354436 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:20.854412924 +0000 UTC m=+160.239857776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.391597 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4r24p"] Feb 18 15:24:20 crc kubenswrapper[4968]: W0218 15:24:20.441472 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddeeab883_52a5_49df_9b71_e3927cf0f2ea.slice/crio-406ebe4186c7b5a8e378089fa5a0d79826d63c40a1fa5eb7f510908fe02af71d WatchSource:0}: Error finding container 406ebe4186c7b5a8e378089fa5a0d79826d63c40a1fa5eb7f510908fe02af71d: Status 404 returned error can't find the container with id 406ebe4186c7b5a8e378089fa5a0d79826d63c40a1fa5eb7f510908fe02af71d Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.454727 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jvwq9"] Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.456034 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:20 crc kubenswrapper[4968]: E0218 15:24:20.456251 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:20.956224992 +0000 UTC m=+160.341669854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.456573 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:20 crc kubenswrapper[4968]: E0218 15:24:20.457009 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:20.956994994 +0000 UTC m=+160.342439856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.558330 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:20 crc kubenswrapper[4968]: E0218 15:24:20.559240 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:21.059217565 +0000 UTC m=+160.444662427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.620508 4968 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.646349 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w8dw5"] Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.660714 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:20 crc kubenswrapper[4968]: E0218 15:24:20.661716 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:21.161648711 +0000 UTC m=+160.547093573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.769870 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:20 crc kubenswrapper[4968]: E0218 15:24:20.770382 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:21.270361701 +0000 UTC m=+160.655806563 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.871779 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:20 crc kubenswrapper[4968]: E0218 15:24:20.872287 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:21.372260362 +0000 UTC m=+160.757705224 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.973335 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:20 crc kubenswrapper[4968]: E0218 15:24:20.973645 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:21.473621866 +0000 UTC m=+160.859066728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.979256 4968 generic.go:334] "Generic (PLEG): container finished" podID="deeab883-52a5-49df-9b71-e3927cf0f2ea" containerID="522783a09d37975d4c568a594147e1c9b7745f88f09af1bf1ee014ae2ff1cfbd" exitCode=0 Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.979587 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4r24p" event={"ID":"deeab883-52a5-49df-9b71-e3927cf0f2ea","Type":"ContainerDied","Data":"522783a09d37975d4c568a594147e1c9b7745f88f09af1bf1ee014ae2ff1cfbd"} Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.979686 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4r24p" event={"ID":"deeab883-52a5-49df-9b71-e3927cf0f2ea","Type":"ContainerStarted","Data":"406ebe4186c7b5a8e378089fa5a0d79826d63c40a1fa5eb7f510908fe02af71d"} Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.981552 4968 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.982337 4968 generic.go:334] "Generic (PLEG): container finished" podID="df2ba163-1935-4600-a6c8-dd73d151ed32" containerID="24a76a6453364bb533f9290a99bed758d12837dc977c57ac28538bc1fbbad3b4" exitCode=0 Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.982406 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" event={"ID":"df2ba163-1935-4600-a6c8-dd73d151ed32","Type":"ContainerDied","Data":"24a76a6453364bb533f9290a99bed758d12837dc977c57ac28538bc1fbbad3b4"} Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.994062 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" event={"ID":"c858a456-78e0-4f36-9a4e-6db413d94d48","Type":"ContainerStarted","Data":"01c06627243a00941455f10749ae9e9af1aeb0521e3918d0e03a19bb3b0375ad"} Feb 18 15:24:20 crc kubenswrapper[4968]: I0218 15:24:20.994125 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" event={"ID":"c858a456-78e0-4f36-9a4e-6db413d94d48","Type":"ContainerStarted","Data":"31f3eae6a3c8382365cebd03f18d29f001f906848fb3beb0f17798bb1c32691a"} Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.003682 4968 generic.go:334] "Generic (PLEG): container finished" podID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" containerID="814d66a6a8897be092bc80d7c0c506fab4f128df8c015dd0bc8a08ad32c941c3" exitCode=0 Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.003825 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w8dw5" event={"ID":"e8871835-fc7c-4c43-a1f2-1450cab78bfe","Type":"ContainerDied","Data":"814d66a6a8897be092bc80d7c0c506fab4f128df8c015dd0bc8a08ad32c941c3"} Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.003863 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w8dw5" event={"ID":"e8871835-fc7c-4c43-a1f2-1450cab78bfe","Type":"ContainerStarted","Data":"58af37126db71a4f2728e6637c43e5b177c3d800c1cf63c4153897fd550a2c20"} Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.008093 4968 generic.go:334] "Generic (PLEG): container finished" podID="190ec188-7d70-42fd-9bca-1d8d6448d005" containerID="98cd99f0745c950c26659dd9a0790123e046c7e1cf9f875ea7fbffb97087c8c7" exitCode=0 Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.008154 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvwq9" event={"ID":"190ec188-7d70-42fd-9bca-1d8d6448d005","Type":"ContainerDied","Data":"98cd99f0745c950c26659dd9a0790123e046c7e1cf9f875ea7fbffb97087c8c7"} Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.008182 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvwq9" event={"ID":"190ec188-7d70-42fd-9bca-1d8d6448d005","Type":"ContainerStarted","Data":"316f8dce16b4b61cde5455bf9a4133431817fff95cbc36d3b99bb7baaa37293c"} Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.011660 4968 generic.go:334] "Generic (PLEG): container finished" podID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" containerID="9c3e4b13a9993481130a68da0b12502fc4d2928ba35bcda91db9b6119ec6e067" exitCode=0 Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.011666 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkb2r" event={"ID":"83659e7b-5ef2-4128-9a9e-4cbe0e332654","Type":"ContainerDied","Data":"9c3e4b13a9993481130a68da0b12502fc4d2928ba35bcda91db9b6119ec6e067"} Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.011782 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkb2r" event={"ID":"83659e7b-5ef2-4128-9a9e-4cbe0e332654","Type":"ContainerStarted","Data":"275972af3588df7723d415062506c3753ddd7a9f6bf1e912efc48ada3a7926a1"} Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.076802 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:21 crc kubenswrapper[4968]: E0218 15:24:21.077184 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:21.577166335 +0000 UTC m=+160.962611197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.085784 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5f4xk"] Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.087659 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.092524 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.096300 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f4xk"] Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.177917 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:21 crc kubenswrapper[4968]: E0218 15:24:21.178053 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:21.678030845 +0000 UTC m=+161.063475707 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.178088 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-catalog-content\") pod \"redhat-marketplace-5f4xk\" (UID: \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\") " pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.178128 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfl8f\" (UniqueName: \"kubernetes.io/projected/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-kube-api-access-sfl8f\") pod \"redhat-marketplace-5f4xk\" (UID: \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\") " pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.178146 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-utilities\") pod \"redhat-marketplace-5f4xk\" (UID: \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\") " pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.178193 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:21 crc kubenswrapper[4968]: E0218 15:24:21.178555 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:21.6785461 +0000 UTC m=+161.063990952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.279123 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.279265 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-utilities\") pod \"redhat-marketplace-5f4xk\" (UID: \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\") " pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:24:21 crc kubenswrapper[4968]: E0218 15:24:21.279339 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 15:24:21.779293047 +0000 UTC m=+161.164737909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.279417 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.279685 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-catalog-content\") pod \"redhat-marketplace-5f4xk\" (UID: \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\") " pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.279736 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-utilities\") pod \"redhat-marketplace-5f4xk\" (UID: \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\") " pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.279789 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfl8f\" (UniqueName: \"kubernetes.io/projected/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-kube-api-access-sfl8f\") pod \"redhat-marketplace-5f4xk\" (UID: \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\") " pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:24:21 crc kubenswrapper[4968]: E0218 15:24:21.280082 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 15:24:21.78006739 +0000 UTC m=+161.165512252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lcx97" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.280609 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-catalog-content\") pod \"redhat-marketplace-5f4xk\" (UID: \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\") " pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.294867 4968 patch_prober.go:28] interesting pod/router-default-5444994796-w9zh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 15:24:21 crc kubenswrapper[4968]: [-]has-synced failed: reason withheld Feb 18 15:24:21 crc kubenswrapper[4968]: [+]process-running ok Feb 18 15:24:21 crc kubenswrapper[4968]: healthz check failed Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.294956 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w9zh9" podUID="1c808375-740d-4fe5-ae91-64f776b28e8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.300633 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfl8f\" (UniqueName: \"kubernetes.io/projected/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-kube-api-access-sfl8f\") pod \"redhat-marketplace-5f4xk\" (UID: \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\") " pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.372017 4968 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-18T15:24:20.620547554Z","Handler":null,"Name":""} Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.375129 4968 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.375172 4968 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.380454 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.384656 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.413285 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.501213 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-289s6"] Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.501467 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.504625 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.508102 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-289s6"] Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.527192 4968 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.527243 4968 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.557243 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lcx97\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.604062 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4179e4cf-dc70-40f8-b710-58614867cfe3-utilities\") pod \"redhat-marketplace-289s6\" (UID: \"4179e4cf-dc70-40f8-b710-58614867cfe3\") " pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.604146 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbtnh\" (UniqueName: \"kubernetes.io/projected/4179e4cf-dc70-40f8-b710-58614867cfe3-kube-api-access-cbtnh\") pod \"redhat-marketplace-289s6\" (UID: \"4179e4cf-dc70-40f8-b710-58614867cfe3\") " pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.604186 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4179e4cf-dc70-40f8-b710-58614867cfe3-catalog-content\") pod \"redhat-marketplace-289s6\" (UID: \"4179e4cf-dc70-40f8-b710-58614867cfe3\") " pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.705281 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4179e4cf-dc70-40f8-b710-58614867cfe3-catalog-content\") pod \"redhat-marketplace-289s6\" (UID: \"4179e4cf-dc70-40f8-b710-58614867cfe3\") " pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.705394 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4179e4cf-dc70-40f8-b710-58614867cfe3-utilities\") pod \"redhat-marketplace-289s6\" (UID: \"4179e4cf-dc70-40f8-b710-58614867cfe3\") " pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.705443 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbtnh\" (UniqueName: \"kubernetes.io/projected/4179e4cf-dc70-40f8-b710-58614867cfe3-kube-api-access-cbtnh\") pod \"redhat-marketplace-289s6\" (UID: \"4179e4cf-dc70-40f8-b710-58614867cfe3\") " pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.707232 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4179e4cf-dc70-40f8-b710-58614867cfe3-catalog-content\") pod \"redhat-marketplace-289s6\" (UID: \"4179e4cf-dc70-40f8-b710-58614867cfe3\") " pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.707527 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4179e4cf-dc70-40f8-b710-58614867cfe3-utilities\") pod \"redhat-marketplace-289s6\" (UID: \"4179e4cf-dc70-40f8-b710-58614867cfe3\") " pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.726704 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbtnh\" (UniqueName: \"kubernetes.io/projected/4179e4cf-dc70-40f8-b710-58614867cfe3-kube-api-access-cbtnh\") pod \"redhat-marketplace-289s6\" (UID: \"4179e4cf-dc70-40f8-b710-58614867cfe3\") " pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.749141 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f4xk"] Feb 18 15:24:21 crc kubenswrapper[4968]: W0218 15:24:21.758496 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48c11632_e38d_4cd3_8ac6_9ef42d55b0e5.slice/crio-fb08b8fcc9b0a35b22b7901248aaf447dee4756881117a6e336eeb334523333a WatchSource:0}: Error finding container fb08b8fcc9b0a35b22b7901248aaf447dee4756881117a6e336eeb334523333a: Status 404 returned error can't find the container with id fb08b8fcc9b0a35b22b7901248aaf447dee4756881117a6e336eeb334523333a Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.800003 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:21 crc kubenswrapper[4968]: I0218 15:24:21.833665 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.048442 4968 generic.go:334] "Generic (PLEG): container finished" podID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" containerID="dac10d240dcacfe95fde80436e862a7d7acb55b2f41c492f593261598039ea32" exitCode=0 Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.050056 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f4xk" event={"ID":"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5","Type":"ContainerDied","Data":"dac10d240dcacfe95fde80436e862a7d7acb55b2f41c492f593261598039ea32"} Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.050149 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f4xk" event={"ID":"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5","Type":"ContainerStarted","Data":"fb08b8fcc9b0a35b22b7901248aaf447dee4756881117a6e336eeb334523333a"} Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.065803 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" event={"ID":"c858a456-78e0-4f36-9a4e-6db413d94d48","Type":"ContainerStarted","Data":"894c0504dab14c26e05722546dca16c6026df76ad2a13b716c098a852c6b0ac3"} Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.125049 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mrbbc"] Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.127567 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.128461 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mrbbc"] Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.130625 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.130833 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-r6dwq" podStartSLOduration=12.130815446 podStartE2EDuration="12.130815446s" podCreationTimestamp="2026-02-18 15:24:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:22.108427479 +0000 UTC m=+161.493872361" watchObservedRunningTime="2026-02-18 15:24:22.130815446 +0000 UTC m=+161.516260308" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.135640 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-289s6"] Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.148978 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lcx97"] Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.224852 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f10928-44af-4829-b9f7-1fe382373a48-utilities\") pod \"redhat-operators-mrbbc\" (UID: \"26f10928-44af-4829-b9f7-1fe382373a48\") " pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.225160 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtcfw\" (UniqueName: \"kubernetes.io/projected/26f10928-44af-4829-b9f7-1fe382373a48-kube-api-access-jtcfw\") pod \"redhat-operators-mrbbc\" (UID: \"26f10928-44af-4829-b9f7-1fe382373a48\") " pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.225220 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f10928-44af-4829-b9f7-1fe382373a48-catalog-content\") pod \"redhat-operators-mrbbc\" (UID: \"26f10928-44af-4829-b9f7-1fe382373a48\") " pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.292558 4968 patch_prober.go:28] interesting pod/router-default-5444994796-w9zh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 15:24:22 crc kubenswrapper[4968]: [-]has-synced failed: reason withheld Feb 18 15:24:22 crc kubenswrapper[4968]: [+]process-running ok Feb 18 15:24:22 crc kubenswrapper[4968]: healthz check failed Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.292631 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w9zh9" podUID="1c808375-740d-4fe5-ae91-64f776b28e8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.326214 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f10928-44af-4829-b9f7-1fe382373a48-utilities\") pod \"redhat-operators-mrbbc\" (UID: \"26f10928-44af-4829-b9f7-1fe382373a48\") " pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.326291 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtcfw\" (UniqueName: \"kubernetes.io/projected/26f10928-44af-4829-b9f7-1fe382373a48-kube-api-access-jtcfw\") pod \"redhat-operators-mrbbc\" (UID: \"26f10928-44af-4829-b9f7-1fe382373a48\") " pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.326322 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f10928-44af-4829-b9f7-1fe382373a48-catalog-content\") pod \"redhat-operators-mrbbc\" (UID: \"26f10928-44af-4829-b9f7-1fe382373a48\") " pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.327316 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f10928-44af-4829-b9f7-1fe382373a48-catalog-content\") pod \"redhat-operators-mrbbc\" (UID: \"26f10928-44af-4829-b9f7-1fe382373a48\") " pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.327781 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f10928-44af-4829-b9f7-1fe382373a48-utilities\") pod \"redhat-operators-mrbbc\" (UID: \"26f10928-44af-4829-b9f7-1fe382373a48\") " pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.354022 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.354261 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtcfw\" (UniqueName: \"kubernetes.io/projected/26f10928-44af-4829-b9f7-1fe382373a48-kube-api-access-jtcfw\") pod \"redhat-operators-mrbbc\" (UID: \"26f10928-44af-4829-b9f7-1fe382373a48\") " pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.428998 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df2ba163-1935-4600-a6c8-dd73d151ed32-config-volume\") pod \"df2ba163-1935-4600-a6c8-dd73d151ed32\" (UID: \"df2ba163-1935-4600-a6c8-dd73d151ed32\") " Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.429703 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df2ba163-1935-4600-a6c8-dd73d151ed32-secret-volume\") pod \"df2ba163-1935-4600-a6c8-dd73d151ed32\" (UID: \"df2ba163-1935-4600-a6c8-dd73d151ed32\") " Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.429778 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrwtn\" (UniqueName: \"kubernetes.io/projected/df2ba163-1935-4600-a6c8-dd73d151ed32-kube-api-access-lrwtn\") pod \"df2ba163-1935-4600-a6c8-dd73d151ed32\" (UID: \"df2ba163-1935-4600-a6c8-dd73d151ed32\") " Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.431127 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df2ba163-1935-4600-a6c8-dd73d151ed32-config-volume" (OuterVolumeSpecName: "config-volume") pod "df2ba163-1935-4600-a6c8-dd73d151ed32" (UID: "df2ba163-1935-4600-a6c8-dd73d151ed32"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.439268 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df2ba163-1935-4600-a6c8-dd73d151ed32-kube-api-access-lrwtn" (OuterVolumeSpecName: "kube-api-access-lrwtn") pod "df2ba163-1935-4600-a6c8-dd73d151ed32" (UID: "df2ba163-1935-4600-a6c8-dd73d151ed32"). InnerVolumeSpecName "kube-api-access-lrwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.441207 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df2ba163-1935-4600-a6c8-dd73d151ed32-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "df2ba163-1935-4600-a6c8-dd73d151ed32" (UID: "df2ba163-1935-4600-a6c8-dd73d151ed32"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.460951 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.478646 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6jpc7"] Feb 18 15:24:22 crc kubenswrapper[4968]: E0218 15:24:22.478986 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df2ba163-1935-4600-a6c8-dd73d151ed32" containerName="collect-profiles" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.479001 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="df2ba163-1935-4600-a6c8-dd73d151ed32" containerName="collect-profiles" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.479158 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="df2ba163-1935-4600-a6c8-dd73d151ed32" containerName="collect-profiles" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.480135 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.495154 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6jpc7"] Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.532515 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vbkk\" (UniqueName: \"kubernetes.io/projected/852b5d5d-bbb3-4123-892d-3cdcad084d67-kube-api-access-7vbkk\") pod \"redhat-operators-6jpc7\" (UID: \"852b5d5d-bbb3-4123-892d-3cdcad084d67\") " pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.532610 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/852b5d5d-bbb3-4123-892d-3cdcad084d67-utilities\") pod \"redhat-operators-6jpc7\" (UID: \"852b5d5d-bbb3-4123-892d-3cdcad084d67\") " pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.532648 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/852b5d5d-bbb3-4123-892d-3cdcad084d67-catalog-content\") pod \"redhat-operators-6jpc7\" (UID: \"852b5d5d-bbb3-4123-892d-3cdcad084d67\") " pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.532902 4968 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df2ba163-1935-4600-a6c8-dd73d151ed32-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.532919 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrwtn\" (UniqueName: \"kubernetes.io/projected/df2ba163-1935-4600-a6c8-dd73d151ed32-kube-api-access-lrwtn\") on node \"crc\" DevicePath \"\"" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.532931 4968 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df2ba163-1935-4600-a6c8-dd73d151ed32-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.634681 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/852b5d5d-bbb3-4123-892d-3cdcad084d67-utilities\") pod \"redhat-operators-6jpc7\" (UID: \"852b5d5d-bbb3-4123-892d-3cdcad084d67\") " pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.634772 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/852b5d5d-bbb3-4123-892d-3cdcad084d67-catalog-content\") pod \"redhat-operators-6jpc7\" (UID: \"852b5d5d-bbb3-4123-892d-3cdcad084d67\") " pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.634872 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vbkk\" (UniqueName: \"kubernetes.io/projected/852b5d5d-bbb3-4123-892d-3cdcad084d67-kube-api-access-7vbkk\") pod \"redhat-operators-6jpc7\" (UID: \"852b5d5d-bbb3-4123-892d-3cdcad084d67\") " pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.635526 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/852b5d5d-bbb3-4123-892d-3cdcad084d67-utilities\") pod \"redhat-operators-6jpc7\" (UID: \"852b5d5d-bbb3-4123-892d-3cdcad084d67\") " pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.635671 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/852b5d5d-bbb3-4123-892d-3cdcad084d67-catalog-content\") pod \"redhat-operators-6jpc7\" (UID: \"852b5d5d-bbb3-4123-892d-3cdcad084d67\") " pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.656977 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vbkk\" (UniqueName: \"kubernetes.io/projected/852b5d5d-bbb3-4123-892d-3cdcad084d67-kube-api-access-7vbkk\") pod \"redhat-operators-6jpc7\" (UID: \"852b5d5d-bbb3-4123-892d-3cdcad084d67\") " pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.760249 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mrbbc"] Feb 18 15:24:22 crc kubenswrapper[4968]: W0218 15:24:22.812393 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26f10928_44af_4829_b9f7_1fe382373a48.slice/crio-1a3d5a4346d54cf5dcb62bdf6eee8f9d1fde9e68130872579c185114fb84b28d WatchSource:0}: Error finding container 1a3d5a4346d54cf5dcb62bdf6eee8f9d1fde9e68130872579c185114fb84b28d: Status 404 returned error can't find the container with id 1a3d5a4346d54cf5dcb62bdf6eee8f9d1fde9e68130872579c185114fb84b28d Feb 18 15:24:22 crc kubenswrapper[4968]: I0218 15:24:22.873828 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.000544 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.001593 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.004139 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.015993 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.017076 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.042143 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9ce53b3-8fd2-48d1-97c9-65e568156feb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b9ce53b3-8fd2-48d1-97c9-65e568156feb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.042227 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b9ce53b3-8fd2-48d1-97c9-65e568156feb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b9ce53b3-8fd2-48d1-97c9-65e568156feb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.112711 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrbbc" event={"ID":"26f10928-44af-4829-b9f7-1fe382373a48","Type":"ContainerStarted","Data":"0784ec62ed1129aaa9288f7d38add433d93111e93fc19a588beacf63718b9a3c"} Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.113169 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrbbc" event={"ID":"26f10928-44af-4829-b9f7-1fe382373a48","Type":"ContainerStarted","Data":"1a3d5a4346d54cf5dcb62bdf6eee8f9d1fde9e68130872579c185114fb84b28d"} Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.120262 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" event={"ID":"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2","Type":"ContainerStarted","Data":"89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078"} Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.120333 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" event={"ID":"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2","Type":"ContainerStarted","Data":"0fee35f6bd2711ee2dfffb4ab1e34b5a89b04cebc5b4b34bf8aca71759cc639f"} Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.120357 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.131244 4968 generic.go:334] "Generic (PLEG): container finished" podID="4179e4cf-dc70-40f8-b710-58614867cfe3" containerID="045b8621674721f917262135e9a2c2d0143d1fc605e0fe17c72a5fe4d2b39467" exitCode=0 Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.131401 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-289s6" event={"ID":"4179e4cf-dc70-40f8-b710-58614867cfe3","Type":"ContainerDied","Data":"045b8621674721f917262135e9a2c2d0143d1fc605e0fe17c72a5fe4d2b39467"} Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.131443 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-289s6" event={"ID":"4179e4cf-dc70-40f8-b710-58614867cfe3","Type":"ContainerStarted","Data":"8a570a25e530959196042d6cdbce19fb9dbee5f39b65c566f15e8ca07c6c858a"} Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.145993 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9ce53b3-8fd2-48d1-97c9-65e568156feb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b9ce53b3-8fd2-48d1-97c9-65e568156feb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.146125 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b9ce53b3-8fd2-48d1-97c9-65e568156feb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b9ce53b3-8fd2-48d1-97c9-65e568156feb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.147913 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b9ce53b3-8fd2-48d1-97c9-65e568156feb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b9ce53b3-8fd2-48d1-97c9-65e568156feb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.150360 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.150454 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-hd6q4" event={"ID":"df2ba163-1935-4600-a6c8-dd73d151ed32","Type":"ContainerDied","Data":"aaf16a5906fa277ca0f32a099e908dd616450afb292fd176f5eb5107a194957c"} Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.150499 4968 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aaf16a5906fa277ca0f32a099e908dd616450afb292fd176f5eb5107a194957c" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.181097 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9ce53b3-8fd2-48d1-97c9-65e568156feb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b9ce53b3-8fd2-48d1-97c9-65e568156feb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.184132 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" podStartSLOduration=142.184098017 podStartE2EDuration="2m22.184098017s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:23.170401735 +0000 UTC m=+162.555846607" watchObservedRunningTime="2026-02-18 15:24:23.184098017 +0000 UTC m=+162.569542879" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.260706 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.288838 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.305086 4968 patch_prober.go:28] interesting pod/router-default-5444994796-w9zh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 15:24:23 crc kubenswrapper[4968]: [-]has-synced failed: reason withheld Feb 18 15:24:23 crc kubenswrapper[4968]: [+]process-running ok Feb 18 15:24:23 crc kubenswrapper[4968]: healthz check failed Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.305187 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w9zh9" podUID="1c808375-740d-4fe5-ae91-64f776b28e8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.326546 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.362254 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6jpc7"] Feb 18 15:24:23 crc kubenswrapper[4968]: W0218 15:24:23.392679 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod852b5d5d_bbb3_4123_892d_3cdcad084d67.slice/crio-0378f205c1401faf94ae70181cc1ae516121a55f15c7ee1b088f07800bc9c42c WatchSource:0}: Error finding container 0378f205c1401faf94ae70181cc1ae516121a55f15c7ee1b088f07800bc9c42c: Status 404 returned error can't find the container with id 0378f205c1401faf94ae70181cc1ae516121a55f15c7ee1b088f07800bc9c42c Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.422965 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.423011 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.427306 4968 patch_prober.go:28] interesting pod/console-f9d7485db-s6qsc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.427393 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-s6qsc" podUID="2c7da228-335d-4f50-83e6-4ae152ed4874" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.444202 4968 patch_prober.go:28] interesting pod/downloads-7954f5f757-td7d6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.444226 4968 patch_prober.go:28] interesting pod/downloads-7954f5f757-td7d6 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.444281 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-td7d6" podUID="5985084a-4c95-41f7-a40a-50550738e31c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.444316 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-td7d6" podUID="5985084a-4c95-41f7-a40a-50550738e31c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.450808 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.452620 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.459191 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.666939 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.676619 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9185b44b-bac7-458b-b4d9-4c389da07c14-metrics-certs\") pod \"network-metrics-daemon-7sk6k\" (UID: \"9185b44b-bac7-458b-b4d9-4c389da07c14\") " pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.676811 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:24:23 crc kubenswrapper[4968]: I0218 15:24:23.765664 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7sk6k" Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.055586 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 18 15:24:24 crc kubenswrapper[4968]: W0218 15:24:24.069451 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podb9ce53b3_8fd2_48d1_97c9_65e568156feb.slice/crio-f68d03ea06de0e3348d3317f82617fb1af151504358a53c52cd80a5bbf5e0652 WatchSource:0}: Error finding container f68d03ea06de0e3348d3317f82617fb1af151504358a53c52cd80a5bbf5e0652: Status 404 returned error can't find the container with id f68d03ea06de0e3348d3317f82617fb1af151504358a53c52cd80a5bbf5e0652 Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.122566 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7sk6k"] Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.169685 4968 generic.go:334] "Generic (PLEG): container finished" podID="26f10928-44af-4829-b9f7-1fe382373a48" containerID="0784ec62ed1129aaa9288f7d38add433d93111e93fc19a588beacf63718b9a3c" exitCode=0 Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.169795 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrbbc" event={"ID":"26f10928-44af-4829-b9f7-1fe382373a48","Type":"ContainerDied","Data":"0784ec62ed1129aaa9288f7d38add433d93111e93fc19a588beacf63718b9a3c"} Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.180341 4968 generic.go:334] "Generic (PLEG): container finished" podID="852b5d5d-bbb3-4123-892d-3cdcad084d67" containerID="2cbee6eb6e9510dc9c41294fa8d2321b06e8378c34179adbe4ebb363c9ac228c" exitCode=0 Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.180628 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jpc7" event={"ID":"852b5d5d-bbb3-4123-892d-3cdcad084d67","Type":"ContainerDied","Data":"2cbee6eb6e9510dc9c41294fa8d2321b06e8378c34179adbe4ebb363c9ac228c"} Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.183601 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jpc7" event={"ID":"852b5d5d-bbb3-4123-892d-3cdcad084d67","Type":"ContainerStarted","Data":"0378f205c1401faf94ae70181cc1ae516121a55f15c7ee1b088f07800bc9c42c"} Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.190633 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b9ce53b3-8fd2-48d1-97c9-65e568156feb","Type":"ContainerStarted","Data":"f68d03ea06de0e3348d3317f82617fb1af151504358a53c52cd80a5bbf5e0652"} Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.211527 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" event={"ID":"9185b44b-bac7-458b-b4d9-4c389da07c14","Type":"ContainerStarted","Data":"0d7e4f1be9afbecf330a0170b614ac1763434f152664d937cad74e0971b5cd0d"} Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.217635 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-whmt5" Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.309067 4968 patch_prober.go:28] interesting pod/router-default-5444994796-w9zh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 15:24:24 crc kubenswrapper[4968]: [-]has-synced failed: reason withheld Feb 18 15:24:24 crc kubenswrapper[4968]: [+]process-running ok Feb 18 15:24:24 crc kubenswrapper[4968]: healthz check failed Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.309180 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w9zh9" podUID="1c808375-740d-4fe5-ae91-64f776b28e8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 15:24:24 crc kubenswrapper[4968]: I0218 15:24:24.628227 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:24:25 crc kubenswrapper[4968]: I0218 15:24:25.269884 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" event={"ID":"9185b44b-bac7-458b-b4d9-4c389da07c14","Type":"ContainerStarted","Data":"dfeb067bf9b7e69d1d7be08d7ae62b77b65ecdd1551386e4ba138ada03232b54"} Feb 18 15:24:25 crc kubenswrapper[4968]: I0218 15:24:25.272633 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b9ce53b3-8fd2-48d1-97c9-65e568156feb","Type":"ContainerStarted","Data":"a515ff5e7fd3fdb8742508211c3791f617e15ac646087fe59a312fe21618547f"} Feb 18 15:24:25 crc kubenswrapper[4968]: I0218 15:24:25.294959 4968 patch_prober.go:28] interesting pod/router-default-5444994796-w9zh9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 15:24:25 crc kubenswrapper[4968]: [-]has-synced failed: reason withheld Feb 18 15:24:25 crc kubenswrapper[4968]: [+]process-running ok Feb 18 15:24:25 crc kubenswrapper[4968]: healthz check failed Feb 18 15:24:25 crc kubenswrapper[4968]: I0218 15:24:25.295389 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w9zh9" podUID="1c808375-740d-4fe5-ae91-64f776b28e8a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 15:24:26 crc kubenswrapper[4968]: I0218 15:24:26.292385 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:26 crc kubenswrapper[4968]: I0218 15:24:26.296430 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-w9zh9" Feb 18 15:24:26 crc kubenswrapper[4968]: I0218 15:24:26.305469 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7sk6k" event={"ID":"9185b44b-bac7-458b-b4d9-4c389da07c14","Type":"ContainerStarted","Data":"bd39fb19b34fa4ebadc4672a6509e71f9acb2d8264b4b708bba59ba0d32b4b75"} Feb 18 15:24:26 crc kubenswrapper[4968]: I0218 15:24:26.317474 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=4.317450101 podStartE2EDuration="4.317450101s" podCreationTimestamp="2026-02-18 15:24:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:25.288420012 +0000 UTC m=+164.673864874" watchObservedRunningTime="2026-02-18 15:24:26.317450101 +0000 UTC m=+165.702894963" Feb 18 15:24:26 crc kubenswrapper[4968]: I0218 15:24:26.323016 4968 generic.go:334] "Generic (PLEG): container finished" podID="b9ce53b3-8fd2-48d1-97c9-65e568156feb" containerID="a515ff5e7fd3fdb8742508211c3791f617e15ac646087fe59a312fe21618547f" exitCode=0 Feb 18 15:24:26 crc kubenswrapper[4968]: I0218 15:24:26.323104 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b9ce53b3-8fd2-48d1-97c9-65e568156feb","Type":"ContainerDied","Data":"a515ff5e7fd3fdb8742508211c3791f617e15ac646087fe59a312fe21618547f"} Feb 18 15:24:26 crc kubenswrapper[4968]: I0218 15:24:26.356028 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-7sk6k" podStartSLOduration=145.355996882 podStartE2EDuration="2m25.355996882s" podCreationTimestamp="2026-02-18 15:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:26.346189354 +0000 UTC m=+165.731634216" watchObservedRunningTime="2026-02-18 15:24:26.355996882 +0000 UTC m=+165.741441734" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.396440 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.397449 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.401832 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.403705 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.408713 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.544888 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbb2f15b-d471-432c-ba83-01f5d75070b5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"cbb2f15b-d471-432c-ba83-01f5d75070b5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.545028 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cbb2f15b-d471-432c-ba83-01f5d75070b5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"cbb2f15b-d471-432c-ba83-01f5d75070b5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.646314 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cbb2f15b-d471-432c-ba83-01f5d75070b5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"cbb2f15b-d471-432c-ba83-01f5d75070b5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.646507 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cbb2f15b-d471-432c-ba83-01f5d75070b5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"cbb2f15b-d471-432c-ba83-01f5d75070b5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.646936 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbb2f15b-d471-432c-ba83-01f5d75070b5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"cbb2f15b-d471-432c-ba83-01f5d75070b5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.695025 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbb2f15b-d471-432c-ba83-01f5d75070b5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"cbb2f15b-d471-432c-ba83-01f5d75070b5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.734591 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.912814 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.954368 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9ce53b3-8fd2-48d1-97c9-65e568156feb-kube-api-access\") pod \"b9ce53b3-8fd2-48d1-97c9-65e568156feb\" (UID: \"b9ce53b3-8fd2-48d1-97c9-65e568156feb\") " Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.954437 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b9ce53b3-8fd2-48d1-97c9-65e568156feb-kubelet-dir\") pod \"b9ce53b3-8fd2-48d1-97c9-65e568156feb\" (UID: \"b9ce53b3-8fd2-48d1-97c9-65e568156feb\") " Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.954845 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9ce53b3-8fd2-48d1-97c9-65e568156feb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b9ce53b3-8fd2-48d1-97c9-65e568156feb" (UID: "b9ce53b3-8fd2-48d1-97c9-65e568156feb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:24:27 crc kubenswrapper[4968]: I0218 15:24:27.976844 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9ce53b3-8fd2-48d1-97c9-65e568156feb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b9ce53b3-8fd2-48d1-97c9-65e568156feb" (UID: "b9ce53b3-8fd2-48d1-97c9-65e568156feb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:24:28 crc kubenswrapper[4968]: I0218 15:24:28.055758 4968 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b9ce53b3-8fd2-48d1-97c9-65e568156feb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 18 15:24:28 crc kubenswrapper[4968]: I0218 15:24:28.055791 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9ce53b3-8fd2-48d1-97c9-65e568156feb-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 15:24:28 crc kubenswrapper[4968]: I0218 15:24:28.195252 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 18 15:24:28 crc kubenswrapper[4968]: W0218 15:24:28.255937 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podcbb2f15b_d471_432c_ba83_01f5d75070b5.slice/crio-3cb939386e1af9a03cf50857e70b520e086d5da83f68d3701369c02be32c6fd5 WatchSource:0}: Error finding container 3cb939386e1af9a03cf50857e70b520e086d5da83f68d3701369c02be32c6fd5: Status 404 returned error can't find the container with id 3cb939386e1af9a03cf50857e70b520e086d5da83f68d3701369c02be32c6fd5 Feb 18 15:24:28 crc kubenswrapper[4968]: I0218 15:24:28.360796 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b9ce53b3-8fd2-48d1-97c9-65e568156feb","Type":"ContainerDied","Data":"f68d03ea06de0e3348d3317f82617fb1af151504358a53c52cd80a5bbf5e0652"} Feb 18 15:24:28 crc kubenswrapper[4968]: I0218 15:24:28.360854 4968 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f68d03ea06de0e3348d3317f82617fb1af151504358a53c52cd80a5bbf5e0652" Feb 18 15:24:28 crc kubenswrapper[4968]: I0218 15:24:28.360880 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 15:24:28 crc kubenswrapper[4968]: I0218 15:24:28.363379 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"cbb2f15b-d471-432c-ba83-01f5d75070b5","Type":"ContainerStarted","Data":"3cb939386e1af9a03cf50857e70b520e086d5da83f68d3701369c02be32c6fd5"} Feb 18 15:24:29 crc kubenswrapper[4968]: I0218 15:24:29.071851 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-d6n9s" Feb 18 15:24:29 crc kubenswrapper[4968]: I0218 15:24:29.393069 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"cbb2f15b-d471-432c-ba83-01f5d75070b5","Type":"ContainerStarted","Data":"429e61e16d933804d3f74cb371c2290e973c14788c9cb44ebf4a0f74dfb94a07"} Feb 18 15:24:29 crc kubenswrapper[4968]: I0218 15:24:29.415059 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.41349357 podStartE2EDuration="2.41349357s" podCreationTimestamp="2026-02-18 15:24:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:24:29.411299446 +0000 UTC m=+168.796744308" watchObservedRunningTime="2026-02-18 15:24:29.41349357 +0000 UTC m=+168.798938432" Feb 18 15:24:30 crc kubenswrapper[4968]: I0218 15:24:30.429167 4968 generic.go:334] "Generic (PLEG): container finished" podID="cbb2f15b-d471-432c-ba83-01f5d75070b5" containerID="429e61e16d933804d3f74cb371c2290e973c14788c9cb44ebf4a0f74dfb94a07" exitCode=0 Feb 18 15:24:30 crc kubenswrapper[4968]: I0218 15:24:30.429659 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"cbb2f15b-d471-432c-ba83-01f5d75070b5","Type":"ContainerDied","Data":"429e61e16d933804d3f74cb371c2290e973c14788c9cb44ebf4a0f74dfb94a07"} Feb 18 15:24:33 crc kubenswrapper[4968]: I0218 15:24:33.423940 4968 patch_prober.go:28] interesting pod/console-f9d7485db-s6qsc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Feb 18 15:24:33 crc kubenswrapper[4968]: I0218 15:24:33.424305 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-s6qsc" podUID="2c7da228-335d-4f50-83e6-4ae152ed4874" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Feb 18 15:24:33 crc kubenswrapper[4968]: I0218 15:24:33.444240 4968 patch_prober.go:28] interesting pod/downloads-7954f5f757-td7d6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 18 15:24:33 crc kubenswrapper[4968]: I0218 15:24:33.444327 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-td7d6" podUID="5985084a-4c95-41f7-a40a-50550738e31c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 18 15:24:33 crc kubenswrapper[4968]: I0218 15:24:33.444425 4968 patch_prober.go:28] interesting pod/downloads-7954f5f757-td7d6 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 18 15:24:33 crc kubenswrapper[4968]: I0218 15:24:33.444528 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-td7d6" podUID="5985084a-4c95-41f7-a40a-50550738e31c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 18 15:24:38 crc kubenswrapper[4968]: I0218 15:24:38.201103 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 15:24:38 crc kubenswrapper[4968]: I0218 15:24:38.221278 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbb2f15b-d471-432c-ba83-01f5d75070b5-kube-api-access\") pod \"cbb2f15b-d471-432c-ba83-01f5d75070b5\" (UID: \"cbb2f15b-d471-432c-ba83-01f5d75070b5\") " Feb 18 15:24:38 crc kubenswrapper[4968]: I0218 15:24:38.221415 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cbb2f15b-d471-432c-ba83-01f5d75070b5-kubelet-dir\") pod \"cbb2f15b-d471-432c-ba83-01f5d75070b5\" (UID: \"cbb2f15b-d471-432c-ba83-01f5d75070b5\") " Feb 18 15:24:38 crc kubenswrapper[4968]: I0218 15:24:38.221717 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbb2f15b-d471-432c-ba83-01f5d75070b5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "cbb2f15b-d471-432c-ba83-01f5d75070b5" (UID: "cbb2f15b-d471-432c-ba83-01f5d75070b5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:24:38 crc kubenswrapper[4968]: I0218 15:24:38.236645 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbb2f15b-d471-432c-ba83-01f5d75070b5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "cbb2f15b-d471-432c-ba83-01f5d75070b5" (UID: "cbb2f15b-d471-432c-ba83-01f5d75070b5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:24:38 crc kubenswrapper[4968]: I0218 15:24:38.322886 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbb2f15b-d471-432c-ba83-01f5d75070b5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 15:24:38 crc kubenswrapper[4968]: I0218 15:24:38.322947 4968 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cbb2f15b-d471-432c-ba83-01f5d75070b5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 18 15:24:38 crc kubenswrapper[4968]: I0218 15:24:38.503681 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"cbb2f15b-d471-432c-ba83-01f5d75070b5","Type":"ContainerDied","Data":"3cb939386e1af9a03cf50857e70b520e086d5da83f68d3701369c02be32c6fd5"} Feb 18 15:24:38 crc kubenswrapper[4968]: I0218 15:24:38.503727 4968 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cb939386e1af9a03cf50857e70b520e086d5da83f68d3701369c02be32c6fd5" Feb 18 15:24:38 crc kubenswrapper[4968]: I0218 15:24:38.503893 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 15:24:39 crc kubenswrapper[4968]: I0218 15:24:39.544601 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 15:24:41 crc kubenswrapper[4968]: I0218 15:24:41.817698 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:24:43 crc kubenswrapper[4968]: I0218 15:24:43.427816 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:43 crc kubenswrapper[4968]: I0218 15:24:43.431651 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:24:43 crc kubenswrapper[4968]: I0218 15:24:43.472460 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-td7d6" Feb 18 15:24:44 crc kubenswrapper[4968]: I0218 15:24:44.368425 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:24:44 crc kubenswrapper[4968]: I0218 15:24:44.368504 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:24:53 crc kubenswrapper[4968]: I0218 15:24:53.279927 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nhsgq" Feb 18 15:24:54 crc kubenswrapper[4968]: E0218 15:24:54.003740 4968 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 18 15:24:54 crc kubenswrapper[4968]: E0218 15:24:54.004172 4968 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kjhbt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-jvwq9_openshift-marketplace(190ec188-7d70-42fd-9bca-1d8d6448d005): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 15:24:54 crc kubenswrapper[4968]: E0218 15:24:54.005450 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-jvwq9" podUID="190ec188-7d70-42fd-9bca-1d8d6448d005" Feb 18 15:24:54 crc kubenswrapper[4968]: E0218 15:24:54.055963 4968 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 18 15:24:54 crc kubenswrapper[4968]: E0218 15:24:54.056427 4968 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lrrhc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hkb2r_openshift-marketplace(83659e7b-5ef2-4128-9a9e-4cbe0e332654): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 15:24:54 crc kubenswrapper[4968]: E0218 15:24:54.057656 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hkb2r" podUID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" Feb 18 15:24:54 crc kubenswrapper[4968]: E0218 15:24:54.062246 4968 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 18 15:24:54 crc kubenswrapper[4968]: E0218 15:24:54.062436 4968 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sfl8f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5f4xk_openshift-marketplace(48c11632-e38d-4cd3-8ac6-9ef42d55b0e5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 15:24:54 crc kubenswrapper[4968]: E0218 15:24:54.063646 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5f4xk" podUID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" Feb 18 15:24:55 crc kubenswrapper[4968]: E0218 15:24:55.561273 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-5f4xk" podUID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" Feb 18 15:24:55 crc kubenswrapper[4968]: E0218 15:24:55.561791 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hkb2r" podUID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" Feb 18 15:24:55 crc kubenswrapper[4968]: E0218 15:24:55.561846 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-jvwq9" podUID="190ec188-7d70-42fd-9bca-1d8d6448d005" Feb 18 15:24:55 crc kubenswrapper[4968]: E0218 15:24:55.660519 4968 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 18 15:24:55 crc kubenswrapper[4968]: E0218 15:24:55.660682 4968 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-58srz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-w8dw5_openshift-marketplace(e8871835-fc7c-4c43-a1f2-1450cab78bfe): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 15:24:55 crc kubenswrapper[4968]: E0218 15:24:55.661966 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-w8dw5" podUID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" Feb 18 15:24:55 crc kubenswrapper[4968]: E0218 15:24:55.664865 4968 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 18 15:24:55 crc kubenswrapper[4968]: E0218 15:24:55.665093 4968 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hrpnd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-4r24p_openshift-marketplace(deeab883-52a5-49df-9b71-e3927cf0f2ea): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 15:24:55 crc kubenswrapper[4968]: E0218 15:24:55.666298 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-4r24p" podUID="deeab883-52a5-49df-9b71-e3927cf0f2ea" Feb 18 15:24:56 crc kubenswrapper[4968]: I0218 15:24:56.614979 4968 generic.go:334] "Generic (PLEG): container finished" podID="4179e4cf-dc70-40f8-b710-58614867cfe3" containerID="f2d6abf1ba37240dc6021e1fd89d644d2b257daba183a7ee4933c11f2a913bcd" exitCode=0 Feb 18 15:24:56 crc kubenswrapper[4968]: I0218 15:24:56.615212 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-289s6" event={"ID":"4179e4cf-dc70-40f8-b710-58614867cfe3","Type":"ContainerDied","Data":"f2d6abf1ba37240dc6021e1fd89d644d2b257daba183a7ee4933c11f2a913bcd"} Feb 18 15:24:56 crc kubenswrapper[4968]: I0218 15:24:56.619965 4968 generic.go:334] "Generic (PLEG): container finished" podID="852b5d5d-bbb3-4123-892d-3cdcad084d67" containerID="dc36206076bca1d13044af217dd15bb0c607767c71a875e8fa7c5ac743d13d80" exitCode=0 Feb 18 15:24:56 crc kubenswrapper[4968]: I0218 15:24:56.620064 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jpc7" event={"ID":"852b5d5d-bbb3-4123-892d-3cdcad084d67","Type":"ContainerDied","Data":"dc36206076bca1d13044af217dd15bb0c607767c71a875e8fa7c5ac743d13d80"} Feb 18 15:24:56 crc kubenswrapper[4968]: I0218 15:24:56.623288 4968 generic.go:334] "Generic (PLEG): container finished" podID="26f10928-44af-4829-b9f7-1fe382373a48" containerID="09a54638221a06794178f0003151e08b499b0459a39c3b6afb46f09037f4f324" exitCode=0 Feb 18 15:24:56 crc kubenswrapper[4968]: I0218 15:24:56.623467 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrbbc" event={"ID":"26f10928-44af-4829-b9f7-1fe382373a48","Type":"ContainerDied","Data":"09a54638221a06794178f0003151e08b499b0459a39c3b6afb46f09037f4f324"} Feb 18 15:24:56 crc kubenswrapper[4968]: E0218 15:24:56.626873 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-4r24p" podUID="deeab883-52a5-49df-9b71-e3927cf0f2ea" Feb 18 15:24:56 crc kubenswrapper[4968]: E0218 15:24:56.626882 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-w8dw5" podUID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" Feb 18 15:24:57 crc kubenswrapper[4968]: I0218 15:24:57.633473 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrbbc" event={"ID":"26f10928-44af-4829-b9f7-1fe382373a48","Type":"ContainerStarted","Data":"8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547"} Feb 18 15:24:57 crc kubenswrapper[4968]: I0218 15:24:57.638309 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jpc7" event={"ID":"852b5d5d-bbb3-4123-892d-3cdcad084d67","Type":"ContainerStarted","Data":"ac2759f012474cd520db41b7f341b558d67b5e68be711ea8cffc0ace7ba85b5a"} Feb 18 15:24:57 crc kubenswrapper[4968]: I0218 15:24:57.641629 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-289s6" event={"ID":"4179e4cf-dc70-40f8-b710-58614867cfe3","Type":"ContainerStarted","Data":"524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f"} Feb 18 15:24:57 crc kubenswrapper[4968]: I0218 15:24:57.655106 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mrbbc" podStartSLOduration=1.595883509 podStartE2EDuration="35.655084682s" podCreationTimestamp="2026-02-18 15:24:22 +0000 UTC" firstStartedPulling="2026-02-18 15:24:23.125121646 +0000 UTC m=+162.510566508" lastFinishedPulling="2026-02-18 15:24:57.184322819 +0000 UTC m=+196.569767681" observedRunningTime="2026-02-18 15:24:57.651255943 +0000 UTC m=+197.036700805" watchObservedRunningTime="2026-02-18 15:24:57.655084682 +0000 UTC m=+197.040529534" Feb 18 15:24:57 crc kubenswrapper[4968]: I0218 15:24:57.689905 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-289s6" podStartSLOduration=2.6012306929999998 podStartE2EDuration="36.689881313s" podCreationTimestamp="2026-02-18 15:24:21 +0000 UTC" firstStartedPulling="2026-02-18 15:24:23.138641983 +0000 UTC m=+162.524086845" lastFinishedPulling="2026-02-18 15:24:57.227292563 +0000 UTC m=+196.612737465" observedRunningTime="2026-02-18 15:24:57.687526106 +0000 UTC m=+197.072970978" watchObservedRunningTime="2026-02-18 15:24:57.689881313 +0000 UTC m=+197.075326175" Feb 18 15:24:57 crc kubenswrapper[4968]: I0218 15:24:57.690719 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6jpc7" podStartSLOduration=2.698513539 podStartE2EDuration="35.690710407s" podCreationTimestamp="2026-02-18 15:24:22 +0000 UTC" firstStartedPulling="2026-02-18 15:24:24.194021345 +0000 UTC m=+163.579466207" lastFinishedPulling="2026-02-18 15:24:57.186218213 +0000 UTC m=+196.571663075" observedRunningTime="2026-02-18 15:24:57.669910484 +0000 UTC m=+197.055355356" watchObservedRunningTime="2026-02-18 15:24:57.690710407 +0000 UTC m=+197.076155269" Feb 18 15:25:01 crc kubenswrapper[4968]: I0218 15:25:01.834368 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:25:01 crc kubenswrapper[4968]: I0218 15:25:01.834688 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:25:02 crc kubenswrapper[4968]: I0218 15:25:02.096391 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:25:02 crc kubenswrapper[4968]: I0218 15:25:02.462151 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:25:02 crc kubenswrapper[4968]: I0218 15:25:02.462237 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:25:02 crc kubenswrapper[4968]: I0218 15:25:02.713499 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:25:02 crc kubenswrapper[4968]: I0218 15:25:02.874019 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:25:02 crc kubenswrapper[4968]: I0218 15:25:02.875108 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:25:03 crc kubenswrapper[4968]: I0218 15:25:03.505308 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mrbbc" podUID="26f10928-44af-4829-b9f7-1fe382373a48" containerName="registry-server" probeResult="failure" output=< Feb 18 15:25:03 crc kubenswrapper[4968]: timeout: failed to connect service ":50051" within 1s Feb 18 15:25:03 crc kubenswrapper[4968]: > Feb 18 15:25:03 crc kubenswrapper[4968]: I0218 15:25:03.838086 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-289s6"] Feb 18 15:25:03 crc kubenswrapper[4968]: I0218 15:25:03.911875 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6jpc7" podUID="852b5d5d-bbb3-4123-892d-3cdcad084d67" containerName="registry-server" probeResult="failure" output=< Feb 18 15:25:03 crc kubenswrapper[4968]: timeout: failed to connect service ":50051" within 1s Feb 18 15:25:03 crc kubenswrapper[4968]: > Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.184067 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 18 15:25:04 crc kubenswrapper[4968]: E0218 15:25:04.184414 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb2f15b-d471-432c-ba83-01f5d75070b5" containerName="pruner" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.184431 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb2f15b-d471-432c-ba83-01f5d75070b5" containerName="pruner" Feb 18 15:25:04 crc kubenswrapper[4968]: E0218 15:25:04.184447 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9ce53b3-8fd2-48d1-97c9-65e568156feb" containerName="pruner" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.184454 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9ce53b3-8fd2-48d1-97c9-65e568156feb" containerName="pruner" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.184551 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9ce53b3-8fd2-48d1-97c9-65e568156feb" containerName="pruner" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.184563 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbb2f15b-d471-432c-ba83-01f5d75070b5" containerName="pruner" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.185094 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.188371 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.189703 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.195162 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.225903 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.226077 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.328621 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.328834 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.328980 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.357365 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.544546 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.679501 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-289s6" podUID="4179e4cf-dc70-40f8-b710-58614867cfe3" containerName="registry-server" containerID="cri-o://524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f" gracePeriod=2 Feb 18 15:25:04 crc kubenswrapper[4968]: I0218 15:25:04.979698 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.129414 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.241590 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbtnh\" (UniqueName: \"kubernetes.io/projected/4179e4cf-dc70-40f8-b710-58614867cfe3-kube-api-access-cbtnh\") pod \"4179e4cf-dc70-40f8-b710-58614867cfe3\" (UID: \"4179e4cf-dc70-40f8-b710-58614867cfe3\") " Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.242091 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4179e4cf-dc70-40f8-b710-58614867cfe3-utilities\") pod \"4179e4cf-dc70-40f8-b710-58614867cfe3\" (UID: \"4179e4cf-dc70-40f8-b710-58614867cfe3\") " Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.242167 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4179e4cf-dc70-40f8-b710-58614867cfe3-catalog-content\") pod \"4179e4cf-dc70-40f8-b710-58614867cfe3\" (UID: \"4179e4cf-dc70-40f8-b710-58614867cfe3\") " Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.243836 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4179e4cf-dc70-40f8-b710-58614867cfe3-utilities" (OuterVolumeSpecName: "utilities") pod "4179e4cf-dc70-40f8-b710-58614867cfe3" (UID: "4179e4cf-dc70-40f8-b710-58614867cfe3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.253984 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4179e4cf-dc70-40f8-b710-58614867cfe3-kube-api-access-cbtnh" (OuterVolumeSpecName: "kube-api-access-cbtnh") pod "4179e4cf-dc70-40f8-b710-58614867cfe3" (UID: "4179e4cf-dc70-40f8-b710-58614867cfe3"). InnerVolumeSpecName "kube-api-access-cbtnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.273552 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4179e4cf-dc70-40f8-b710-58614867cfe3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4179e4cf-dc70-40f8-b710-58614867cfe3" (UID: "4179e4cf-dc70-40f8-b710-58614867cfe3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.343431 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbtnh\" (UniqueName: \"kubernetes.io/projected/4179e4cf-dc70-40f8-b710-58614867cfe3-kube-api-access-cbtnh\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.343486 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4179e4cf-dc70-40f8-b710-58614867cfe3-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.343499 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4179e4cf-dc70-40f8-b710-58614867cfe3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.686995 4968 generic.go:334] "Generic (PLEG): container finished" podID="4179e4cf-dc70-40f8-b710-58614867cfe3" containerID="524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f" exitCode=0 Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.687060 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-289s6" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.687057 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-289s6" event={"ID":"4179e4cf-dc70-40f8-b710-58614867cfe3","Type":"ContainerDied","Data":"524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f"} Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.687102 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-289s6" event={"ID":"4179e4cf-dc70-40f8-b710-58614867cfe3","Type":"ContainerDied","Data":"8a570a25e530959196042d6cdbce19fb9dbee5f39b65c566f15e8ca07c6c858a"} Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.687124 4968 scope.go:117] "RemoveContainer" containerID="524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.689481 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7","Type":"ContainerStarted","Data":"ebc7e038a64408e5295eccffaca7f3f12a49aba050c353b5becc1aac80e7d175"} Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.689507 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7","Type":"ContainerStarted","Data":"716a6e12019eb124ba76d66e219c9e6b840e3b1d27e73c01f024364ff77dd4d5"} Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.713294 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.713138893 podStartE2EDuration="1.713138893s" podCreationTimestamp="2026-02-18 15:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:25:05.71127519 +0000 UTC m=+205.096720052" watchObservedRunningTime="2026-02-18 15:25:05.713138893 +0000 UTC m=+205.098583755" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.729804 4968 scope.go:117] "RemoveContainer" containerID="f2d6abf1ba37240dc6021e1fd89d644d2b257daba183a7ee4933c11f2a913bcd" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.730948 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-289s6"] Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.735484 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-289s6"] Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.757863 4968 scope.go:117] "RemoveContainer" containerID="045b8621674721f917262135e9a2c2d0143d1fc605e0fe17c72a5fe4d2b39467" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.773175 4968 scope.go:117] "RemoveContainer" containerID="524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f" Feb 18 15:25:05 crc kubenswrapper[4968]: E0218 15:25:05.774850 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f\": container with ID starting with 524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f not found: ID does not exist" containerID="524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.774895 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f"} err="failed to get container status \"524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f\": rpc error: code = NotFound desc = could not find container \"524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f\": container with ID starting with 524634f1b01383588407dd42804a7836859ae48d089cb52d0a145246bd9a834f not found: ID does not exist" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.774946 4968 scope.go:117] "RemoveContainer" containerID="f2d6abf1ba37240dc6021e1fd89d644d2b257daba183a7ee4933c11f2a913bcd" Feb 18 15:25:05 crc kubenswrapper[4968]: E0218 15:25:05.775449 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2d6abf1ba37240dc6021e1fd89d644d2b257daba183a7ee4933c11f2a913bcd\": container with ID starting with f2d6abf1ba37240dc6021e1fd89d644d2b257daba183a7ee4933c11f2a913bcd not found: ID does not exist" containerID="f2d6abf1ba37240dc6021e1fd89d644d2b257daba183a7ee4933c11f2a913bcd" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.775476 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2d6abf1ba37240dc6021e1fd89d644d2b257daba183a7ee4933c11f2a913bcd"} err="failed to get container status \"f2d6abf1ba37240dc6021e1fd89d644d2b257daba183a7ee4933c11f2a913bcd\": rpc error: code = NotFound desc = could not find container \"f2d6abf1ba37240dc6021e1fd89d644d2b257daba183a7ee4933c11f2a913bcd\": container with ID starting with f2d6abf1ba37240dc6021e1fd89d644d2b257daba183a7ee4933c11f2a913bcd not found: ID does not exist" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.775491 4968 scope.go:117] "RemoveContainer" containerID="045b8621674721f917262135e9a2c2d0143d1fc605e0fe17c72a5fe4d2b39467" Feb 18 15:25:05 crc kubenswrapper[4968]: E0218 15:25:05.775763 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"045b8621674721f917262135e9a2c2d0143d1fc605e0fe17c72a5fe4d2b39467\": container with ID starting with 045b8621674721f917262135e9a2c2d0143d1fc605e0fe17c72a5fe4d2b39467 not found: ID does not exist" containerID="045b8621674721f917262135e9a2c2d0143d1fc605e0fe17c72a5fe4d2b39467" Feb 18 15:25:05 crc kubenswrapper[4968]: I0218 15:25:05.775787 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"045b8621674721f917262135e9a2c2d0143d1fc605e0fe17c72a5fe4d2b39467"} err="failed to get container status \"045b8621674721f917262135e9a2c2d0143d1fc605e0fe17c72a5fe4d2b39467\": rpc error: code = NotFound desc = could not find container \"045b8621674721f917262135e9a2c2d0143d1fc605e0fe17c72a5fe4d2b39467\": container with ID starting with 045b8621674721f917262135e9a2c2d0143d1fc605e0fe17c72a5fe4d2b39467 not found: ID does not exist" Feb 18 15:25:06 crc kubenswrapper[4968]: I0218 15:25:06.700406 4968 generic.go:334] "Generic (PLEG): container finished" podID="9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7" containerID="ebc7e038a64408e5295eccffaca7f3f12a49aba050c353b5becc1aac80e7d175" exitCode=0 Feb 18 15:25:06 crc kubenswrapper[4968]: I0218 15:25:06.700460 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7","Type":"ContainerDied","Data":"ebc7e038a64408e5295eccffaca7f3f12a49aba050c353b5becc1aac80e7d175"} Feb 18 15:25:07 crc kubenswrapper[4968]: I0218 15:25:07.240959 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4179e4cf-dc70-40f8-b710-58614867cfe3" path="/var/lib/kubelet/pods/4179e4cf-dc70-40f8-b710-58614867cfe3/volumes" Feb 18 15:25:08 crc kubenswrapper[4968]: I0218 15:25:08.000199 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 15:25:08 crc kubenswrapper[4968]: I0218 15:25:08.185883 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7-kube-api-access\") pod \"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7\" (UID: \"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7\") " Feb 18 15:25:08 crc kubenswrapper[4968]: I0218 15:25:08.185984 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7-kubelet-dir\") pod \"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7\" (UID: \"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7\") " Feb 18 15:25:08 crc kubenswrapper[4968]: I0218 15:25:08.186660 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7" (UID: "9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:25:08 crc kubenswrapper[4968]: I0218 15:25:08.193336 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7" (UID: "9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:25:08 crc kubenswrapper[4968]: I0218 15:25:08.287951 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:08 crc kubenswrapper[4968]: I0218 15:25:08.287998 4968 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:08 crc kubenswrapper[4968]: I0218 15:25:08.715927 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7","Type":"ContainerDied","Data":"716a6e12019eb124ba76d66e219c9e6b840e3b1d27e73c01f024364ff77dd4d5"} Feb 18 15:25:08 crc kubenswrapper[4968]: I0218 15:25:08.715988 4968 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="716a6e12019eb124ba76d66e219c9e6b840e3b1d27e73c01f024364ff77dd4d5" Feb 18 15:25:08 crc kubenswrapper[4968]: I0218 15:25:08.716028 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.183228 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 18 15:25:11 crc kubenswrapper[4968]: E0218 15:25:11.183671 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4179e4cf-dc70-40f8-b710-58614867cfe3" containerName="extract-utilities" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.183682 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="4179e4cf-dc70-40f8-b710-58614867cfe3" containerName="extract-utilities" Feb 18 15:25:11 crc kubenswrapper[4968]: E0218 15:25:11.183690 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4179e4cf-dc70-40f8-b710-58614867cfe3" containerName="registry-server" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.183696 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="4179e4cf-dc70-40f8-b710-58614867cfe3" containerName="registry-server" Feb 18 15:25:11 crc kubenswrapper[4968]: E0218 15:25:11.183708 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7" containerName="pruner" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.183715 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7" containerName="pruner" Feb 18 15:25:11 crc kubenswrapper[4968]: E0218 15:25:11.183731 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4179e4cf-dc70-40f8-b710-58614867cfe3" containerName="extract-content" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.183737 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="4179e4cf-dc70-40f8-b710-58614867cfe3" containerName="extract-content" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.183845 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="4179e4cf-dc70-40f8-b710-58614867cfe3" containerName="registry-server" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.183860 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ff07c7f-a4dd-4e4d-970f-b4f4fb11cab7" containerName="pruner" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.184195 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.186361 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.187102 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.192183 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.351391 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3802ffab-07cd-449d-b890-4b25e1571d57-kube-api-access\") pod \"installer-9-crc\" (UID: \"3802ffab-07cd-449d-b890-4b25e1571d57\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.351629 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3802ffab-07cd-449d-b890-4b25e1571d57-var-lock\") pod \"installer-9-crc\" (UID: \"3802ffab-07cd-449d-b890-4b25e1571d57\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.351769 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3802ffab-07cd-449d-b890-4b25e1571d57-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3802ffab-07cd-449d-b890-4b25e1571d57\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.453516 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3802ffab-07cd-449d-b890-4b25e1571d57-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3802ffab-07cd-449d-b890-4b25e1571d57\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.453657 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3802ffab-07cd-449d-b890-4b25e1571d57-kube-api-access\") pod \"installer-9-crc\" (UID: \"3802ffab-07cd-449d-b890-4b25e1571d57\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.453701 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3802ffab-07cd-449d-b890-4b25e1571d57-var-lock\") pod \"installer-9-crc\" (UID: \"3802ffab-07cd-449d-b890-4b25e1571d57\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.453714 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3802ffab-07cd-449d-b890-4b25e1571d57-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3802ffab-07cd-449d-b890-4b25e1571d57\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.453826 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3802ffab-07cd-449d-b890-4b25e1571d57-var-lock\") pod \"installer-9-crc\" (UID: \"3802ffab-07cd-449d-b890-4b25e1571d57\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.472594 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3802ffab-07cd-449d-b890-4b25e1571d57-kube-api-access\") pod \"installer-9-crc\" (UID: \"3802ffab-07cd-449d-b890-4b25e1571d57\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.510202 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.755626 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f4xk" event={"ID":"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5","Type":"ContainerStarted","Data":"4835c97fbcf6aee48a483c3f83bdde3894eaf7239be9351d9178fb0ff9989865"} Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.759701 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w8dw5" event={"ID":"e8871835-fc7c-4c43-a1f2-1450cab78bfe","Type":"ContainerStarted","Data":"935286d4faff7d33380ef28b643c77a36620937ab9174a0afc6063026570cc91"} Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.762873 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkb2r" event={"ID":"83659e7b-5ef2-4128-9a9e-4cbe0e332654","Type":"ContainerStarted","Data":"370174d7c6b8a5d86f2b039b1495819498c0ad720d184695a683709b27914337"} Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.783909 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 18 15:25:11 crc kubenswrapper[4968]: W0218 15:25:11.795771 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod3802ffab_07cd_449d_b890_4b25e1571d57.slice/crio-fdccf15b8510ae11ff2115bc2b34b870b427f8b2f3e2ef7ab458389d50cac66b WatchSource:0}: Error finding container fdccf15b8510ae11ff2115bc2b34b870b427f8b2f3e2ef7ab458389d50cac66b: Status 404 returned error can't find the container with id fdccf15b8510ae11ff2115bc2b34b870b427f8b2f3e2ef7ab458389d50cac66b Feb 18 15:25:11 crc kubenswrapper[4968]: I0218 15:25:11.820641 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8g9pk"] Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.512527 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.558908 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.770931 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3802ffab-07cd-449d-b890-4b25e1571d57","Type":"ContainerStarted","Data":"27e37b717dd4715056fc11464eb5592f2fea76da2964ababc2718efe908d1399"} Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.770992 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3802ffab-07cd-449d-b890-4b25e1571d57","Type":"ContainerStarted","Data":"fdccf15b8510ae11ff2115bc2b34b870b427f8b2f3e2ef7ab458389d50cac66b"} Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.776757 4968 generic.go:334] "Generic (PLEG): container finished" podID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" containerID="370174d7c6b8a5d86f2b039b1495819498c0ad720d184695a683709b27914337" exitCode=0 Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.776821 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkb2r" event={"ID":"83659e7b-5ef2-4128-9a9e-4cbe0e332654","Type":"ContainerDied","Data":"370174d7c6b8a5d86f2b039b1495819498c0ad720d184695a683709b27914337"} Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.778650 4968 generic.go:334] "Generic (PLEG): container finished" podID="190ec188-7d70-42fd-9bca-1d8d6448d005" containerID="ecef50a1e963bb4514ec06c8b27826bc8203f02e6a75d489c639bd58293e109c" exitCode=0 Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.778727 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvwq9" event={"ID":"190ec188-7d70-42fd-9bca-1d8d6448d005","Type":"ContainerDied","Data":"ecef50a1e963bb4514ec06c8b27826bc8203f02e6a75d489c639bd58293e109c"} Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.783439 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4r24p" event={"ID":"deeab883-52a5-49df-9b71-e3927cf0f2ea","Type":"ContainerStarted","Data":"27a3f09f40054a3959682f11d2b0d7f37b99e3bb61270998dff67cfadb2a96e4"} Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.785685 4968 generic.go:334] "Generic (PLEG): container finished" podID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" containerID="4835c97fbcf6aee48a483c3f83bdde3894eaf7239be9351d9178fb0ff9989865" exitCode=0 Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.785776 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f4xk" event={"ID":"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5","Type":"ContainerDied","Data":"4835c97fbcf6aee48a483c3f83bdde3894eaf7239be9351d9178fb0ff9989865"} Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.789043 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.78902495 podStartE2EDuration="1.78902495s" podCreationTimestamp="2026-02-18 15:25:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:25:12.78800074 +0000 UTC m=+212.173445602" watchObservedRunningTime="2026-02-18 15:25:12.78902495 +0000 UTC m=+212.174469812" Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.797830 4968 generic.go:334] "Generic (PLEG): container finished" podID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" containerID="935286d4faff7d33380ef28b643c77a36620937ab9174a0afc6063026570cc91" exitCode=0 Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.797945 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w8dw5" event={"ID":"e8871835-fc7c-4c43-a1f2-1450cab78bfe","Type":"ContainerDied","Data":"935286d4faff7d33380ef28b643c77a36620937ab9174a0afc6063026570cc91"} Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.909951 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:25:12 crc kubenswrapper[4968]: I0218 15:25:12.957731 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:25:13 crc kubenswrapper[4968]: I0218 15:25:13.807512 4968 generic.go:334] "Generic (PLEG): container finished" podID="deeab883-52a5-49df-9b71-e3927cf0f2ea" containerID="27a3f09f40054a3959682f11d2b0d7f37b99e3bb61270998dff67cfadb2a96e4" exitCode=0 Feb 18 15:25:13 crc kubenswrapper[4968]: I0218 15:25:13.807574 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4r24p" event={"ID":"deeab883-52a5-49df-9b71-e3927cf0f2ea","Type":"ContainerDied","Data":"27a3f09f40054a3959682f11d2b0d7f37b99e3bb61270998dff67cfadb2a96e4"} Feb 18 15:25:14 crc kubenswrapper[4968]: I0218 15:25:14.367358 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:25:14 crc kubenswrapper[4968]: I0218 15:25:14.367458 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:25:14 crc kubenswrapper[4968]: I0218 15:25:14.367532 4968 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:25:14 crc kubenswrapper[4968]: I0218 15:25:14.368686 4968 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5"} pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:25:14 crc kubenswrapper[4968]: I0218 15:25:14.368936 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" containerID="cri-o://940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5" gracePeriod=600 Feb 18 15:25:14 crc kubenswrapper[4968]: I0218 15:25:14.816889 4968 generic.go:334] "Generic (PLEG): container finished" podID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerID="940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5" exitCode=0 Feb 18 15:25:14 crc kubenswrapper[4968]: I0218 15:25:14.816938 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerDied","Data":"940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5"} Feb 18 15:25:14 crc kubenswrapper[4968]: I0218 15:25:14.820818 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvwq9" event={"ID":"190ec188-7d70-42fd-9bca-1d8d6448d005","Type":"ContainerStarted","Data":"71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559"} Feb 18 15:25:15 crc kubenswrapper[4968]: I0218 15:25:15.828875 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkb2r" event={"ID":"83659e7b-5ef2-4128-9a9e-4cbe0e332654","Type":"ContainerStarted","Data":"f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a"} Feb 18 15:25:15 crc kubenswrapper[4968]: I0218 15:25:15.832713 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerStarted","Data":"5d6c8b54ab16252892873fa8a0af6e5bfc9fe64986de280b4035d7ef2a1e29f8"} Feb 18 15:25:15 crc kubenswrapper[4968]: I0218 15:25:15.836254 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4r24p" event={"ID":"deeab883-52a5-49df-9b71-e3927cf0f2ea","Type":"ContainerStarted","Data":"53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455"} Feb 18 15:25:15 crc kubenswrapper[4968]: I0218 15:25:15.838206 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f4xk" event={"ID":"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5","Type":"ContainerStarted","Data":"86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e"} Feb 18 15:25:15 crc kubenswrapper[4968]: I0218 15:25:15.841136 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w8dw5" event={"ID":"e8871835-fc7c-4c43-a1f2-1450cab78bfe","Type":"ContainerStarted","Data":"6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd"} Feb 18 15:25:15 crc kubenswrapper[4968]: I0218 15:25:15.851028 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hkb2r" podStartSLOduration=3.475487241 podStartE2EDuration="57.851008812s" podCreationTimestamp="2026-02-18 15:24:18 +0000 UTC" firstStartedPulling="2026-02-18 15:24:21.014049583 +0000 UTC m=+160.399494445" lastFinishedPulling="2026-02-18 15:25:15.389571134 +0000 UTC m=+214.775016016" observedRunningTime="2026-02-18 15:25:15.848967043 +0000 UTC m=+215.234411925" watchObservedRunningTime="2026-02-18 15:25:15.851008812 +0000 UTC m=+215.236453674" Feb 18 15:25:15 crc kubenswrapper[4968]: I0218 15:25:15.943110 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w8dw5" podStartSLOduration=2.387439366 podStartE2EDuration="56.943078065s" podCreationTimestamp="2026-02-18 15:24:19 +0000 UTC" firstStartedPulling="2026-02-18 15:24:21.006960215 +0000 UTC m=+160.392405077" lastFinishedPulling="2026-02-18 15:25:15.562598914 +0000 UTC m=+214.948043776" observedRunningTime="2026-02-18 15:25:15.93201952 +0000 UTC m=+215.317464392" watchObservedRunningTime="2026-02-18 15:25:15.943078065 +0000 UTC m=+215.328522927" Feb 18 15:25:15 crc kubenswrapper[4968]: I0218 15:25:15.978575 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4r24p" podStartSLOduration=2.423416119 podStartE2EDuration="56.978549586s" podCreationTimestamp="2026-02-18 15:24:19 +0000 UTC" firstStartedPulling="2026-02-18 15:24:20.981114816 +0000 UTC m=+160.366559678" lastFinishedPulling="2026-02-18 15:25:15.536248283 +0000 UTC m=+214.921693145" observedRunningTime="2026-02-18 15:25:15.973860032 +0000 UTC m=+215.359304884" watchObservedRunningTime="2026-02-18 15:25:15.978549586 +0000 UTC m=+215.363994448" Feb 18 15:25:15 crc kubenswrapper[4968]: I0218 15:25:15.996792 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5f4xk" podStartSLOduration=1.651856831 podStartE2EDuration="54.996774915s" podCreationTimestamp="2026-02-18 15:24:21 +0000 UTC" firstStartedPulling="2026-02-18 15:24:22.060302876 +0000 UTC m=+161.445747738" lastFinishedPulling="2026-02-18 15:25:15.40522095 +0000 UTC m=+214.790665822" observedRunningTime="2026-02-18 15:25:15.996081065 +0000 UTC m=+215.381525927" watchObservedRunningTime="2026-02-18 15:25:15.996774915 +0000 UTC m=+215.382219777" Feb 18 15:25:16 crc kubenswrapper[4968]: I0218 15:25:16.019949 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jvwq9" podStartSLOduration=3.759403459 podStartE2EDuration="57.019929835s" podCreationTimestamp="2026-02-18 15:24:19 +0000 UTC" firstStartedPulling="2026-02-18 15:24:21.010266192 +0000 UTC m=+160.395711054" lastFinishedPulling="2026-02-18 15:25:14.270792578 +0000 UTC m=+213.656237430" observedRunningTime="2026-02-18 15:25:16.017464894 +0000 UTC m=+215.402909756" watchObservedRunningTime="2026-02-18 15:25:16.019929835 +0000 UTC m=+215.405374707" Feb 18 15:25:16 crc kubenswrapper[4968]: I0218 15:25:16.640024 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6jpc7"] Feb 18 15:25:16 crc kubenswrapper[4968]: I0218 15:25:16.640580 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6jpc7" podUID="852b5d5d-bbb3-4123-892d-3cdcad084d67" containerName="registry-server" containerID="cri-o://ac2759f012474cd520db41b7f341b558d67b5e68be711ea8cffc0ace7ba85b5a" gracePeriod=2 Feb 18 15:25:16 crc kubenswrapper[4968]: I0218 15:25:16.848541 4968 generic.go:334] "Generic (PLEG): container finished" podID="852b5d5d-bbb3-4123-892d-3cdcad084d67" containerID="ac2759f012474cd520db41b7f341b558d67b5e68be711ea8cffc0ace7ba85b5a" exitCode=0 Feb 18 15:25:16 crc kubenswrapper[4968]: I0218 15:25:16.848640 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jpc7" event={"ID":"852b5d5d-bbb3-4123-892d-3cdcad084d67","Type":"ContainerDied","Data":"ac2759f012474cd520db41b7f341b558d67b5e68be711ea8cffc0ace7ba85b5a"} Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.057232 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.232400 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/852b5d5d-bbb3-4123-892d-3cdcad084d67-catalog-content\") pod \"852b5d5d-bbb3-4123-892d-3cdcad084d67\" (UID: \"852b5d5d-bbb3-4123-892d-3cdcad084d67\") " Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.232531 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/852b5d5d-bbb3-4123-892d-3cdcad084d67-utilities\") pod \"852b5d5d-bbb3-4123-892d-3cdcad084d67\" (UID: \"852b5d5d-bbb3-4123-892d-3cdcad084d67\") " Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.232569 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vbkk\" (UniqueName: \"kubernetes.io/projected/852b5d5d-bbb3-4123-892d-3cdcad084d67-kube-api-access-7vbkk\") pod \"852b5d5d-bbb3-4123-892d-3cdcad084d67\" (UID: \"852b5d5d-bbb3-4123-892d-3cdcad084d67\") " Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.233310 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/852b5d5d-bbb3-4123-892d-3cdcad084d67-utilities" (OuterVolumeSpecName: "utilities") pod "852b5d5d-bbb3-4123-892d-3cdcad084d67" (UID: "852b5d5d-bbb3-4123-892d-3cdcad084d67"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.241438 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/852b5d5d-bbb3-4123-892d-3cdcad084d67-kube-api-access-7vbkk" (OuterVolumeSpecName: "kube-api-access-7vbkk") pod "852b5d5d-bbb3-4123-892d-3cdcad084d67" (UID: "852b5d5d-bbb3-4123-892d-3cdcad084d67"). InnerVolumeSpecName "kube-api-access-7vbkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.334055 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/852b5d5d-bbb3-4123-892d-3cdcad084d67-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.334117 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vbkk\" (UniqueName: \"kubernetes.io/projected/852b5d5d-bbb3-4123-892d-3cdcad084d67-kube-api-access-7vbkk\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.392584 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/852b5d5d-bbb3-4123-892d-3cdcad084d67-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "852b5d5d-bbb3-4123-892d-3cdcad084d67" (UID: "852b5d5d-bbb3-4123-892d-3cdcad084d67"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.435914 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/852b5d5d-bbb3-4123-892d-3cdcad084d67-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.856629 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jpc7" event={"ID":"852b5d5d-bbb3-4123-892d-3cdcad084d67","Type":"ContainerDied","Data":"0378f205c1401faf94ae70181cc1ae516121a55f15c7ee1b088f07800bc9c42c"} Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.856701 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6jpc7" Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.856978 4968 scope.go:117] "RemoveContainer" containerID="ac2759f012474cd520db41b7f341b558d67b5e68be711ea8cffc0ace7ba85b5a" Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.892137 4968 scope.go:117] "RemoveContainer" containerID="dc36206076bca1d13044af217dd15bb0c607767c71a875e8fa7c5ac743d13d80" Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.892942 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6jpc7"] Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.897654 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6jpc7"] Feb 18 15:25:17 crc kubenswrapper[4968]: I0218 15:25:17.913141 4968 scope.go:117] "RemoveContainer" containerID="2cbee6eb6e9510dc9c41294fa8d2321b06e8378c34179adbe4ebb363c9ac228c" Feb 18 15:25:19 crc kubenswrapper[4968]: I0218 15:25:19.236911 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="852b5d5d-bbb3-4123-892d-3cdcad084d67" path="/var/lib/kubelet/pods/852b5d5d-bbb3-4123-892d-3cdcad084d67/volumes" Feb 18 15:25:19 crc kubenswrapper[4968]: I0218 15:25:19.237600 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:25:19 crc kubenswrapper[4968]: I0218 15:25:19.237639 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:25:19 crc kubenswrapper[4968]: I0218 15:25:19.296103 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:25:19 crc kubenswrapper[4968]: I0218 15:25:19.707397 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:25:19 crc kubenswrapper[4968]: I0218 15:25:19.707453 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:25:19 crc kubenswrapper[4968]: I0218 15:25:19.746441 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:25:19 crc kubenswrapper[4968]: I0218 15:25:19.747097 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:25:19 crc kubenswrapper[4968]: I0218 15:25:19.748067 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:25:19 crc kubenswrapper[4968]: I0218 15:25:19.793841 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:25:19 crc kubenswrapper[4968]: I0218 15:25:19.911173 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:25:20 crc kubenswrapper[4968]: I0218 15:25:20.112949 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:25:20 crc kubenswrapper[4968]: I0218 15:25:20.113706 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:25:20 crc kubenswrapper[4968]: I0218 15:25:20.153225 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:25:20 crc kubenswrapper[4968]: I0218 15:25:20.929579 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:25:21 crc kubenswrapper[4968]: I0218 15:25:21.413976 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:25:21 crc kubenswrapper[4968]: I0218 15:25:21.414039 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:25:21 crc kubenswrapper[4968]: I0218 15:25:21.455275 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:25:21 crc kubenswrapper[4968]: I0218 15:25:21.925818 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.047549 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w8dw5"] Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.048537 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w8dw5" podUID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" containerName="registry-server" containerID="cri-o://6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd" gracePeriod=2 Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.563013 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.718909 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8871835-fc7c-4c43-a1f2-1450cab78bfe-utilities\") pod \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\" (UID: \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\") " Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.719110 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8871835-fc7c-4c43-a1f2-1450cab78bfe-catalog-content\") pod \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\" (UID: \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\") " Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.719147 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58srz\" (UniqueName: \"kubernetes.io/projected/e8871835-fc7c-4c43-a1f2-1450cab78bfe-kube-api-access-58srz\") pod \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\" (UID: \"e8871835-fc7c-4c43-a1f2-1450cab78bfe\") " Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.719792 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8871835-fc7c-4c43-a1f2-1450cab78bfe-utilities" (OuterVolumeSpecName: "utilities") pod "e8871835-fc7c-4c43-a1f2-1450cab78bfe" (UID: "e8871835-fc7c-4c43-a1f2-1450cab78bfe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.724860 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8871835-fc7c-4c43-a1f2-1450cab78bfe-kube-api-access-58srz" (OuterVolumeSpecName: "kube-api-access-58srz") pod "e8871835-fc7c-4c43-a1f2-1450cab78bfe" (UID: "e8871835-fc7c-4c43-a1f2-1450cab78bfe"). InnerVolumeSpecName "kube-api-access-58srz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.794983 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8871835-fc7c-4c43-a1f2-1450cab78bfe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8871835-fc7c-4c43-a1f2-1450cab78bfe" (UID: "e8871835-fc7c-4c43-a1f2-1450cab78bfe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.820215 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8871835-fc7c-4c43-a1f2-1450cab78bfe-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.820261 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58srz\" (UniqueName: \"kubernetes.io/projected/e8871835-fc7c-4c43-a1f2-1450cab78bfe-kube-api-access-58srz\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.820276 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8871835-fc7c-4c43-a1f2-1450cab78bfe-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.893920 4968 generic.go:334] "Generic (PLEG): container finished" podID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" containerID="6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd" exitCode=0 Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.893979 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w8dw5" event={"ID":"e8871835-fc7c-4c43-a1f2-1450cab78bfe","Type":"ContainerDied","Data":"6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd"} Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.894014 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w8dw5" event={"ID":"e8871835-fc7c-4c43-a1f2-1450cab78bfe","Type":"ContainerDied","Data":"58af37126db71a4f2728e6637c43e5b177c3d800c1cf63c4153897fd550a2c20"} Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.894040 4968 scope.go:117] "RemoveContainer" containerID="6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.894200 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w8dw5" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.914270 4968 scope.go:117] "RemoveContainer" containerID="935286d4faff7d33380ef28b643c77a36620937ab9174a0afc6063026570cc91" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.936227 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w8dw5"] Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.937361 4968 scope.go:117] "RemoveContainer" containerID="814d66a6a8897be092bc80d7c0c506fab4f128df8c015dd0bc8a08ad32c941c3" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.941915 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w8dw5"] Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.964063 4968 scope.go:117] "RemoveContainer" containerID="6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd" Feb 18 15:25:23 crc kubenswrapper[4968]: E0218 15:25:23.964719 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd\": container with ID starting with 6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd not found: ID does not exist" containerID="6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.964777 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd"} err="failed to get container status \"6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd\": rpc error: code = NotFound desc = could not find container \"6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd\": container with ID starting with 6a91703d82afdcfadb5860e2f135a59e713ffd6799e48ff59834947114cf3acd not found: ID does not exist" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.964806 4968 scope.go:117] "RemoveContainer" containerID="935286d4faff7d33380ef28b643c77a36620937ab9174a0afc6063026570cc91" Feb 18 15:25:23 crc kubenswrapper[4968]: E0218 15:25:23.965217 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"935286d4faff7d33380ef28b643c77a36620937ab9174a0afc6063026570cc91\": container with ID starting with 935286d4faff7d33380ef28b643c77a36620937ab9174a0afc6063026570cc91 not found: ID does not exist" containerID="935286d4faff7d33380ef28b643c77a36620937ab9174a0afc6063026570cc91" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.965254 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"935286d4faff7d33380ef28b643c77a36620937ab9174a0afc6063026570cc91"} err="failed to get container status \"935286d4faff7d33380ef28b643c77a36620937ab9174a0afc6063026570cc91\": rpc error: code = NotFound desc = could not find container \"935286d4faff7d33380ef28b643c77a36620937ab9174a0afc6063026570cc91\": container with ID starting with 935286d4faff7d33380ef28b643c77a36620937ab9174a0afc6063026570cc91 not found: ID does not exist" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.965280 4968 scope.go:117] "RemoveContainer" containerID="814d66a6a8897be092bc80d7c0c506fab4f128df8c015dd0bc8a08ad32c941c3" Feb 18 15:25:23 crc kubenswrapper[4968]: E0218 15:25:23.965722 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"814d66a6a8897be092bc80d7c0c506fab4f128df8c015dd0bc8a08ad32c941c3\": container with ID starting with 814d66a6a8897be092bc80d7c0c506fab4f128df8c015dd0bc8a08ad32c941c3 not found: ID does not exist" containerID="814d66a6a8897be092bc80d7c0c506fab4f128df8c015dd0bc8a08ad32c941c3" Feb 18 15:25:23 crc kubenswrapper[4968]: I0218 15:25:23.965757 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"814d66a6a8897be092bc80d7c0c506fab4f128df8c015dd0bc8a08ad32c941c3"} err="failed to get container status \"814d66a6a8897be092bc80d7c0c506fab4f128df8c015dd0bc8a08ad32c941c3\": rpc error: code = NotFound desc = could not find container \"814d66a6a8897be092bc80d7c0c506fab4f128df8c015dd0bc8a08ad32c941c3\": container with ID starting with 814d66a6a8897be092bc80d7c0c506fab4f128df8c015dd0bc8a08ad32c941c3 not found: ID does not exist" Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.038247 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jvwq9"] Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.038512 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jvwq9" podUID="190ec188-7d70-42fd-9bca-1d8d6448d005" containerName="registry-server" containerID="cri-o://71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559" gracePeriod=2 Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.894096 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.904879 4968 generic.go:334] "Generic (PLEG): container finished" podID="190ec188-7d70-42fd-9bca-1d8d6448d005" containerID="71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559" exitCode=0 Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.904942 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvwq9" event={"ID":"190ec188-7d70-42fd-9bca-1d8d6448d005","Type":"ContainerDied","Data":"71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559"} Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.904981 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvwq9" event={"ID":"190ec188-7d70-42fd-9bca-1d8d6448d005","Type":"ContainerDied","Data":"316f8dce16b4b61cde5455bf9a4133431817fff95cbc36d3b99bb7baaa37293c"} Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.905004 4968 scope.go:117] "RemoveContainer" containerID="71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559" Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.905170 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvwq9" Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.939031 4968 scope.go:117] "RemoveContainer" containerID="ecef50a1e963bb4514ec06c8b27826bc8203f02e6a75d489c639bd58293e109c" Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.965393 4968 scope.go:117] "RemoveContainer" containerID="98cd99f0745c950c26659dd9a0790123e046c7e1cf9f875ea7fbffb97087c8c7" Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.980692 4968 scope.go:117] "RemoveContainer" containerID="71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559" Feb 18 15:25:24 crc kubenswrapper[4968]: E0218 15:25:24.981170 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559\": container with ID starting with 71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559 not found: ID does not exist" containerID="71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559" Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.981230 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559"} err="failed to get container status \"71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559\": rpc error: code = NotFound desc = could not find container \"71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559\": container with ID starting with 71b7e8190a4a2c3b9e1fd0ac5ce785e386c362becf501ce47e6ef25f1cee9559 not found: ID does not exist" Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.981269 4968 scope.go:117] "RemoveContainer" containerID="ecef50a1e963bb4514ec06c8b27826bc8203f02e6a75d489c639bd58293e109c" Feb 18 15:25:24 crc kubenswrapper[4968]: E0218 15:25:24.981573 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecef50a1e963bb4514ec06c8b27826bc8203f02e6a75d489c639bd58293e109c\": container with ID starting with ecef50a1e963bb4514ec06c8b27826bc8203f02e6a75d489c639bd58293e109c not found: ID does not exist" containerID="ecef50a1e963bb4514ec06c8b27826bc8203f02e6a75d489c639bd58293e109c" Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.981607 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecef50a1e963bb4514ec06c8b27826bc8203f02e6a75d489c639bd58293e109c"} err="failed to get container status \"ecef50a1e963bb4514ec06c8b27826bc8203f02e6a75d489c639bd58293e109c\": rpc error: code = NotFound desc = could not find container \"ecef50a1e963bb4514ec06c8b27826bc8203f02e6a75d489c639bd58293e109c\": container with ID starting with ecef50a1e963bb4514ec06c8b27826bc8203f02e6a75d489c639bd58293e109c not found: ID does not exist" Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.981629 4968 scope.go:117] "RemoveContainer" containerID="98cd99f0745c950c26659dd9a0790123e046c7e1cf9f875ea7fbffb97087c8c7" Feb 18 15:25:24 crc kubenswrapper[4968]: E0218 15:25:24.981892 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98cd99f0745c950c26659dd9a0790123e046c7e1cf9f875ea7fbffb97087c8c7\": container with ID starting with 98cd99f0745c950c26659dd9a0790123e046c7e1cf9f875ea7fbffb97087c8c7 not found: ID does not exist" containerID="98cd99f0745c950c26659dd9a0790123e046c7e1cf9f875ea7fbffb97087c8c7" Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.981928 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98cd99f0745c950c26659dd9a0790123e046c7e1cf9f875ea7fbffb97087c8c7"} err="failed to get container status \"98cd99f0745c950c26659dd9a0790123e046c7e1cf9f875ea7fbffb97087c8c7\": rpc error: code = NotFound desc = could not find container \"98cd99f0745c950c26659dd9a0790123e046c7e1cf9f875ea7fbffb97087c8c7\": container with ID starting with 98cd99f0745c950c26659dd9a0790123e046c7e1cf9f875ea7fbffb97087c8c7 not found: ID does not exist" Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.995931 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/190ec188-7d70-42fd-9bca-1d8d6448d005-catalog-content\") pod \"190ec188-7d70-42fd-9bca-1d8d6448d005\" (UID: \"190ec188-7d70-42fd-9bca-1d8d6448d005\") " Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.995991 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjhbt\" (UniqueName: \"kubernetes.io/projected/190ec188-7d70-42fd-9bca-1d8d6448d005-kube-api-access-kjhbt\") pod \"190ec188-7d70-42fd-9bca-1d8d6448d005\" (UID: \"190ec188-7d70-42fd-9bca-1d8d6448d005\") " Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.996160 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/190ec188-7d70-42fd-9bca-1d8d6448d005-utilities\") pod \"190ec188-7d70-42fd-9bca-1d8d6448d005\" (UID: \"190ec188-7d70-42fd-9bca-1d8d6448d005\") " Feb 18 15:25:24 crc kubenswrapper[4968]: I0218 15:25:24.997044 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/190ec188-7d70-42fd-9bca-1d8d6448d005-utilities" (OuterVolumeSpecName: "utilities") pod "190ec188-7d70-42fd-9bca-1d8d6448d005" (UID: "190ec188-7d70-42fd-9bca-1d8d6448d005"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:25:25 crc kubenswrapper[4968]: I0218 15:25:25.004559 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/190ec188-7d70-42fd-9bca-1d8d6448d005-kube-api-access-kjhbt" (OuterVolumeSpecName: "kube-api-access-kjhbt") pod "190ec188-7d70-42fd-9bca-1d8d6448d005" (UID: "190ec188-7d70-42fd-9bca-1d8d6448d005"). InnerVolumeSpecName "kube-api-access-kjhbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:25:25 crc kubenswrapper[4968]: I0218 15:25:25.048424 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/190ec188-7d70-42fd-9bca-1d8d6448d005-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "190ec188-7d70-42fd-9bca-1d8d6448d005" (UID: "190ec188-7d70-42fd-9bca-1d8d6448d005"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:25:25 crc kubenswrapper[4968]: I0218 15:25:25.097848 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjhbt\" (UniqueName: \"kubernetes.io/projected/190ec188-7d70-42fd-9bca-1d8d6448d005-kube-api-access-kjhbt\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:25 crc kubenswrapper[4968]: I0218 15:25:25.097882 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/190ec188-7d70-42fd-9bca-1d8d6448d005-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:25 crc kubenswrapper[4968]: I0218 15:25:25.097891 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/190ec188-7d70-42fd-9bca-1d8d6448d005-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:25 crc kubenswrapper[4968]: I0218 15:25:25.247775 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" path="/var/lib/kubelet/pods/e8871835-fc7c-4c43-a1f2-1450cab78bfe/volumes" Feb 18 15:25:25 crc kubenswrapper[4968]: I0218 15:25:25.248394 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jvwq9"] Feb 18 15:25:25 crc kubenswrapper[4968]: I0218 15:25:25.248427 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jvwq9"] Feb 18 15:25:27 crc kubenswrapper[4968]: I0218 15:25:27.237295 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="190ec188-7d70-42fd-9bca-1d8d6448d005" path="/var/lib/kubelet/pods/190ec188-7d70-42fd-9bca-1d8d6448d005/volumes" Feb 18 15:25:29 crc kubenswrapper[4968]: I0218 15:25:29.276324 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:25:29 crc kubenswrapper[4968]: I0218 15:25:29.787189 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:25:36 crc kubenswrapper[4968]: I0218 15:25:36.873205 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" podUID="57795f27-c9c3-4f38-a5da-364e2488f343" containerName="oauth-openshift" containerID="cri-o://9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5" gracePeriod=15 Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.251200 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274430 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57795f27-c9c3-4f38-a5da-364e2488f343-audit-dir\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274525 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-cliconfig\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274560 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-trusted-ca-bundle\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274586 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-969jw\" (UniqueName: \"kubernetes.io/projected/57795f27-c9c3-4f38-a5da-364e2488f343-kube-api-access-969jw\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274666 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-error\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274700 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-provider-selection\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274773 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-session\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274809 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-router-certs\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274837 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-ocp-branding-template\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274883 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-idp-0-file-data\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274908 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-audit-policies\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274937 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-serving-cert\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274962 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-login\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.274996 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-service-ca\") pod \"57795f27-c9c3-4f38-a5da-364e2488f343\" (UID: \"57795f27-c9c3-4f38-a5da-364e2488f343\") " Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.279325 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57795f27-c9c3-4f38-a5da-364e2488f343-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.280378 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.281214 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.283779 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.285765 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.286447 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-58d58b5989-mxd5q"] Feb 18 15:25:37 crc kubenswrapper[4968]: E0218 15:25:37.286829 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="190ec188-7d70-42fd-9bca-1d8d6448d005" containerName="extract-content" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.286976 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="190ec188-7d70-42fd-9bca-1d8d6448d005" containerName="extract-content" Feb 18 15:25:37 crc kubenswrapper[4968]: E0218 15:25:37.287063 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="852b5d5d-bbb3-4123-892d-3cdcad084d67" containerName="extract-utilities" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.287131 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="852b5d5d-bbb3-4123-892d-3cdcad084d67" containerName="extract-utilities" Feb 18 15:25:37 crc kubenswrapper[4968]: E0218 15:25:37.287204 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="852b5d5d-bbb3-4123-892d-3cdcad084d67" containerName="extract-content" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.287269 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="852b5d5d-bbb3-4123-892d-3cdcad084d67" containerName="extract-content" Feb 18 15:25:37 crc kubenswrapper[4968]: E0218 15:25:37.287334 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="190ec188-7d70-42fd-9bca-1d8d6448d005" containerName="extract-utilities" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.287403 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="190ec188-7d70-42fd-9bca-1d8d6448d005" containerName="extract-utilities" Feb 18 15:25:37 crc kubenswrapper[4968]: E0218 15:25:37.287465 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" containerName="extract-utilities" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.287522 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" containerName="extract-utilities" Feb 18 15:25:37 crc kubenswrapper[4968]: E0218 15:25:37.287584 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="190ec188-7d70-42fd-9bca-1d8d6448d005" containerName="registry-server" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.287657 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="190ec188-7d70-42fd-9bca-1d8d6448d005" containerName="registry-server" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.287219 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: E0218 15:25:37.287805 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="852b5d5d-bbb3-4123-892d-3cdcad084d67" containerName="registry-server" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.287880 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="852b5d5d-bbb3-4123-892d-3cdcad084d67" containerName="registry-server" Feb 18 15:25:37 crc kubenswrapper[4968]: E0218 15:25:37.287953 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" containerName="extract-content" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.288014 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" containerName="extract-content" Feb 18 15:25:37 crc kubenswrapper[4968]: E0218 15:25:37.288082 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57795f27-c9c3-4f38-a5da-364e2488f343" containerName="oauth-openshift" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.288141 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="57795f27-c9c3-4f38-a5da-364e2488f343" containerName="oauth-openshift" Feb 18 15:25:37 crc kubenswrapper[4968]: E0218 15:25:37.288221 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" containerName="registry-server" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.288361 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" containerName="registry-server" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.288152 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.288833 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="190ec188-7d70-42fd-9bca-1d8d6448d005" containerName="registry-server" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.288935 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="57795f27-c9c3-4f38-a5da-364e2488f343" containerName="oauth-openshift" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.289015 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="852b5d5d-bbb3-4123-892d-3cdcad084d67" containerName="registry-server" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.289730 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8871835-fc7c-4c43-a1f2-1450cab78bfe" containerName="registry-server" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.289927 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57795f27-c9c3-4f38-a5da-364e2488f343-kube-api-access-969jw" (OuterVolumeSpecName: "kube-api-access-969jw") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "kube-api-access-969jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.290496 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.290682 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.296036 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.296098 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-58d58b5989-mxd5q"] Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.296329 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.301546 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.304858 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.307915 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "57795f27-c9c3-4f38-a5da-364e2488f343" (UID: "57795f27-c9c3-4f38-a5da-364e2488f343"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376085 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376155 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376187 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-service-ca\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376220 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376260 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376321 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-user-template-login\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376365 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376405 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-session\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376464 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-router-certs\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376513 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5mpl\" (UniqueName: \"kubernetes.io/projected/c7c03800-2271-4033-9b27-ef7fcb4956f7-kube-api-access-c5mpl\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376589 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-user-template-error\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376620 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376648 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c7c03800-2271-4033-9b27-ef7fcb4956f7-audit-policies\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376677 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7c03800-2271-4033-9b27-ef7fcb4956f7-audit-dir\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376739 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376777 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376791 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376806 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376819 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376832 4968 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376844 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376857 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376869 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376883 4968 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57795f27-c9c3-4f38-a5da-364e2488f343-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376895 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376907 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376922 4968 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/57795f27-c9c3-4f38-a5da-364e2488f343-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.376935 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-969jw\" (UniqueName: \"kubernetes.io/projected/57795f27-c9c3-4f38-a5da-364e2488f343-kube-api-access-969jw\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.478743 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-user-template-error\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.478875 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.478960 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c7c03800-2271-4033-9b27-ef7fcb4956f7-audit-policies\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.479006 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7c03800-2271-4033-9b27-ef7fcb4956f7-audit-dir\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.479053 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.479092 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.479128 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-service-ca\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.479163 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.479170 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7c03800-2271-4033-9b27-ef7fcb4956f7-audit-dir\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.479210 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.479253 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-user-template-login\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.479305 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.479389 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-session\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.479459 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-router-certs\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.479498 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5mpl\" (UniqueName: \"kubernetes.io/projected/c7c03800-2271-4033-9b27-ef7fcb4956f7-kube-api-access-c5mpl\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.480247 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c7c03800-2271-4033-9b27-ef7fcb4956f7-audit-policies\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.480954 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-service-ca\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.487563 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.487914 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.490808 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-router-certs\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.491523 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.492489 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-user-template-error\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.492823 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.493325 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-session\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.493437 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.494533 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-user-template-login\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.498632 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c03800-2271-4033-9b27-ef7fcb4956f7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.509558 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5mpl\" (UniqueName: \"kubernetes.io/projected/c7c03800-2271-4033-9b27-ef7fcb4956f7-kube-api-access-c5mpl\") pod \"oauth-openshift-58d58b5989-mxd5q\" (UID: \"c7c03800-2271-4033-9b27-ef7fcb4956f7\") " pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.638983 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.854127 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-58d58b5989-mxd5q"] Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.978608 4968 generic.go:334] "Generic (PLEG): container finished" podID="57795f27-c9c3-4f38-a5da-364e2488f343" containerID="9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5" exitCode=0 Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.978690 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.978716 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" event={"ID":"57795f27-c9c3-4f38-a5da-364e2488f343","Type":"ContainerDied","Data":"9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5"} Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.978830 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8g9pk" event={"ID":"57795f27-c9c3-4f38-a5da-364e2488f343","Type":"ContainerDied","Data":"231f552c329f92d7e13f1ce9560950b542e0b97058f4b23d2263e8f4427e6ffd"} Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.978849 4968 scope.go:117] "RemoveContainer" containerID="9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5" Feb 18 15:25:37 crc kubenswrapper[4968]: I0218 15:25:37.979739 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" event={"ID":"c7c03800-2271-4033-9b27-ef7fcb4956f7","Type":"ContainerStarted","Data":"dfc9ecd8b5fac315b1ded202849691ee9a185f1371df031334ac406f0404b617"} Feb 18 15:25:38 crc kubenswrapper[4968]: I0218 15:25:38.026249 4968 scope.go:117] "RemoveContainer" containerID="9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5" Feb 18 15:25:38 crc kubenswrapper[4968]: I0218 15:25:38.032579 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8g9pk"] Feb 18 15:25:38 crc kubenswrapper[4968]: E0218 15:25:38.032776 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5\": container with ID starting with 9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5 not found: ID does not exist" containerID="9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5" Feb 18 15:25:38 crc kubenswrapper[4968]: I0218 15:25:38.032809 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5"} err="failed to get container status \"9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5\": rpc error: code = NotFound desc = could not find container \"9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5\": container with ID starting with 9f04b548825d584118beacd013da59126d15bf7255668ac5d3d0736346eac0b5 not found: ID does not exist" Feb 18 15:25:38 crc kubenswrapper[4968]: I0218 15:25:38.037343 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8g9pk"] Feb 18 15:25:38 crc kubenswrapper[4968]: I0218 15:25:38.987056 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" event={"ID":"c7c03800-2271-4033-9b27-ef7fcb4956f7","Type":"ContainerStarted","Data":"5f9488023752d168f1313cfc9e457167ecf322bb9b68c636c3f647bd1ddd50f4"} Feb 18 15:25:38 crc kubenswrapper[4968]: I0218 15:25:38.987321 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:38 crc kubenswrapper[4968]: I0218 15:25:38.993315 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" Feb 18 15:25:39 crc kubenswrapper[4968]: I0218 15:25:39.020100 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-58d58b5989-mxd5q" podStartSLOduration=28.020085478 podStartE2EDuration="28.020085478s" podCreationTimestamp="2026-02-18 15:25:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:25:39.014976192 +0000 UTC m=+238.400421054" watchObservedRunningTime="2026-02-18 15:25:39.020085478 +0000 UTC m=+238.405530340" Feb 18 15:25:39 crc kubenswrapper[4968]: I0218 15:25:39.237361 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57795f27-c9c3-4f38-a5da-364e2488f343" path="/var/lib/kubelet/pods/57795f27-c9c3-4f38-a5da-364e2488f343/volumes" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.989305 4968 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.990691 4968 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.991023 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32" gracePeriod=15 Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.991081 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.991092 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4" gracePeriod=15 Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.991122 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd" gracePeriod=15 Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.991249 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91" gracePeriod=15 Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.991155 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd" gracePeriod=15 Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.993543 4968 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 15:25:49 crc kubenswrapper[4968]: E0218 15:25:49.993872 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.993890 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 18 15:25:49 crc kubenswrapper[4968]: E0218 15:25:49.993902 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.993911 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 18 15:25:49 crc kubenswrapper[4968]: E0218 15:25:49.993929 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.993937 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 18 15:25:49 crc kubenswrapper[4968]: E0218 15:25:49.993955 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.993963 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 15:25:49 crc kubenswrapper[4968]: E0218 15:25:49.993975 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.993984 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 18 15:25:49 crc kubenswrapper[4968]: E0218 15:25:49.993994 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.994003 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 15:25:49 crc kubenswrapper[4968]: E0218 15:25:49.994016 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.994024 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.994142 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.994165 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.994178 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.994189 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.994198 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 15:25:49 crc kubenswrapper[4968]: I0218 15:25:49.994207 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.040393 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.040476 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.040536 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.040566 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.040630 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.040660 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.040688 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.040735 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142059 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142145 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142176 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142225 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142233 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142265 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142291 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142304 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142247 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142235 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142358 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142441 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142471 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142499 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142521 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.142618 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.767900 4968 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 18 15:25:50 crc kubenswrapper[4968]: I0218 15:25:50.767992 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.072812 4968 generic.go:334] "Generic (PLEG): container finished" podID="3802ffab-07cd-449d-b890-4b25e1571d57" containerID="27e37b717dd4715056fc11464eb5592f2fea76da2964ababc2718efe908d1399" exitCode=0 Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.072908 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3802ffab-07cd-449d-b890-4b25e1571d57","Type":"ContainerDied","Data":"27e37b717dd4715056fc11464eb5592f2fea76da2964ababc2718efe908d1399"} Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.073878 4968 status_manager.go:851] "Failed to get status for pod" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.074201 4968 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.075658 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.077090 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.077926 4968 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4" exitCode=0 Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.077951 4968 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91" exitCode=0 Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.077962 4968 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd" exitCode=0 Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.077971 4968 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd" exitCode=2 Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.078003 4968 scope.go:117] "RemoveContainer" containerID="a7fb99e1b019b0ca9a2098d01834646666658392f9bc173a7a4a226ec288fb97" Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.232463 4968 status_manager.go:851] "Failed to get status for pod" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.232682 4968 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:51 crc kubenswrapper[4968]: E0218 15:25:51.734841 4968 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:51 crc kubenswrapper[4968]: E0218 15:25:51.735561 4968 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:51 crc kubenswrapper[4968]: E0218 15:25:51.736162 4968 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:51 crc kubenswrapper[4968]: E0218 15:25:51.736490 4968 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:51 crc kubenswrapper[4968]: E0218 15:25:51.736821 4968 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:51 crc kubenswrapper[4968]: I0218 15:25:51.736854 4968 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 18 15:25:51 crc kubenswrapper[4968]: E0218 15:25:51.737122 4968 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.151:6443: connect: connection refused" interval="200ms" Feb 18 15:25:51 crc kubenswrapper[4968]: E0218 15:25:51.938256 4968 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.151:6443: connect: connection refused" interval="400ms" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.107147 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 18 15:25:52 crc kubenswrapper[4968]: E0218 15:25:52.339266 4968 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.151:6443: connect: connection refused" interval="800ms" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.394018 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.395049 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.395693 4968 status_manager.go:851] "Failed to get status for pod" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.396053 4968 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.409440 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.410280 4968 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.411163 4968 status_manager.go:851] "Failed to get status for pod" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.571920 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3802ffab-07cd-449d-b890-4b25e1571d57-kubelet-dir\") pod \"3802ffab-07cd-449d-b890-4b25e1571d57\" (UID: \"3802ffab-07cd-449d-b890-4b25e1571d57\") " Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.571985 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.572019 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.572036 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.572081 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3802ffab-07cd-449d-b890-4b25e1571d57-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3802ffab-07cd-449d-b890-4b25e1571d57" (UID: "3802ffab-07cd-449d-b890-4b25e1571d57"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.572097 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.572136 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.572190 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.572150 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3802ffab-07cd-449d-b890-4b25e1571d57-kube-api-access\") pod \"3802ffab-07cd-449d-b890-4b25e1571d57\" (UID: \"3802ffab-07cd-449d-b890-4b25e1571d57\") " Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.572317 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3802ffab-07cd-449d-b890-4b25e1571d57-var-lock\") pod \"3802ffab-07cd-449d-b890-4b25e1571d57\" (UID: \"3802ffab-07cd-449d-b890-4b25e1571d57\") " Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.572423 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3802ffab-07cd-449d-b890-4b25e1571d57-var-lock" (OuterVolumeSpecName: "var-lock") pod "3802ffab-07cd-449d-b890-4b25e1571d57" (UID: "3802ffab-07cd-449d-b890-4b25e1571d57"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.572982 4968 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.573045 4968 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.573065 4968 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3802ffab-07cd-449d-b890-4b25e1571d57-var-lock\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.573112 4968 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3802ffab-07cd-449d-b890-4b25e1571d57-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.573129 4968 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.580927 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3802ffab-07cd-449d-b890-4b25e1571d57-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3802ffab-07cd-449d-b890-4b25e1571d57" (UID: "3802ffab-07cd-449d-b890-4b25e1571d57"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:25:52 crc kubenswrapper[4968]: I0218 15:25:52.674472 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3802ffab-07cd-449d-b890-4b25e1571d57-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.121855 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3802ffab-07cd-449d-b890-4b25e1571d57","Type":"ContainerDied","Data":"fdccf15b8510ae11ff2115bc2b34b870b427f8b2f3e2ef7ab458389d50cac66b"} Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.122466 4968 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdccf15b8510ae11ff2115bc2b34b870b427f8b2f3e2ef7ab458389d50cac66b" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.122695 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.124163 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.125070 4968 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32" exitCode=0 Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.125130 4968 scope.go:117] "RemoveContainer" containerID="a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.125166 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:25:53 crc kubenswrapper[4968]: E0218 15:25:53.140064 4968 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.151:6443: connect: connection refused" interval="1.6s" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.143128 4968 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.143498 4968 status_manager.go:851] "Failed to get status for pod" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.143732 4968 status_manager.go:851] "Failed to get status for pod" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.144389 4968 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.144979 4968 scope.go:117] "RemoveContainer" containerID="38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.158926 4968 scope.go:117] "RemoveContainer" containerID="d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.173650 4968 scope.go:117] "RemoveContainer" containerID="37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.188029 4968 scope.go:117] "RemoveContainer" containerID="df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.203326 4968 scope.go:117] "RemoveContainer" containerID="a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.226010 4968 scope.go:117] "RemoveContainer" containerID="a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4" Feb 18 15:25:53 crc kubenswrapper[4968]: E0218 15:25:53.226560 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\": container with ID starting with a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4 not found: ID does not exist" containerID="a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.226594 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4"} err="failed to get container status \"a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\": rpc error: code = NotFound desc = could not find container \"a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4\": container with ID starting with a03628da0928eaff528c864f4b24449e749b0c63619160f8441df31f146bf3b4 not found: ID does not exist" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.226635 4968 scope.go:117] "RemoveContainer" containerID="38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91" Feb 18 15:25:53 crc kubenswrapper[4968]: E0218 15:25:53.226938 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\": container with ID starting with 38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91 not found: ID does not exist" containerID="38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.226965 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91"} err="failed to get container status \"38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\": rpc error: code = NotFound desc = could not find container \"38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91\": container with ID starting with 38c4ce3d5a76742826a1be9c1f116033a323597fdb586ad9ec472d0a6030fd91 not found: ID does not exist" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.226984 4968 scope.go:117] "RemoveContainer" containerID="d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd" Feb 18 15:25:53 crc kubenswrapper[4968]: E0218 15:25:53.227230 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\": container with ID starting with d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd not found: ID does not exist" containerID="d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.227255 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd"} err="failed to get container status \"d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\": rpc error: code = NotFound desc = could not find container \"d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd\": container with ID starting with d9c56038c79e8e2a8ffde9df837595c8a9983893f3f5da64ab03cee94c083bbd not found: ID does not exist" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.227271 4968 scope.go:117] "RemoveContainer" containerID="37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd" Feb 18 15:25:53 crc kubenswrapper[4968]: E0218 15:25:53.227621 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\": container with ID starting with 37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd not found: ID does not exist" containerID="37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.227640 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd"} err="failed to get container status \"37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\": rpc error: code = NotFound desc = could not find container \"37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd\": container with ID starting with 37cfa509c3a021cae48f709e73142b4c15b84f3e7dd8292017a4670954dbcbbd not found: ID does not exist" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.227652 4968 scope.go:117] "RemoveContainer" containerID="df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32" Feb 18 15:25:53 crc kubenswrapper[4968]: E0218 15:25:53.228474 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\": container with ID starting with df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32 not found: ID does not exist" containerID="df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.228505 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32"} err="failed to get container status \"df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\": rpc error: code = NotFound desc = could not find container \"df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32\": container with ID starting with df9303f1e7f2b89cf1890ddbbb108f0a00fcfc26a44395a7a9c76340c4db8c32 not found: ID does not exist" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.228522 4968 scope.go:117] "RemoveContainer" containerID="a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb" Feb 18 15:25:53 crc kubenswrapper[4968]: E0218 15:25:53.229701 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\": container with ID starting with a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb not found: ID does not exist" containerID="a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.229726 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb"} err="failed to get container status \"a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\": rpc error: code = NotFound desc = could not find container \"a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb\": container with ID starting with a1420f02e0d1736276f1d1c84348ab752730d87cbecde7cb6cb5eee749f277fb not found: ID does not exist" Feb 18 15:25:53 crc kubenswrapper[4968]: I0218 15:25:53.237259 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 18 15:25:54 crc kubenswrapper[4968]: E0218 15:25:54.740955 4968 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.151:6443: connect: connection refused" interval="3.2s" Feb 18 15:25:55 crc kubenswrapper[4968]: E0218 15:25:55.023569 4968 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.151:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:55 crc kubenswrapper[4968]: I0218 15:25:55.024154 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:55 crc kubenswrapper[4968]: E0218 15:25:55.060155 4968 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.151:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189560b41862744c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-18 15:25:55.059643468 +0000 UTC m=+254.445088370,LastTimestamp:2026-02-18 15:25:55.059643468 +0000 UTC m=+254.445088370,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 18 15:25:55 crc kubenswrapper[4968]: I0218 15:25:55.139852 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"6239b6124bb2d9f1c7785393b54918695f256296a0ac8c1b30fe8580a126ba3f"} Feb 18 15:25:56 crc kubenswrapper[4968]: I0218 15:25:56.144993 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"82eaa531149f6c4f6cbe9b4c5816880d95efa5bacf479bdbfe8c55bf3a452b38"} Feb 18 15:25:56 crc kubenswrapper[4968]: I0218 15:25:56.145600 4968 status_manager.go:851] "Failed to get status for pod" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:25:56 crc kubenswrapper[4968]: E0218 15:25:56.145662 4968 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.151:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:57 crc kubenswrapper[4968]: E0218 15:25:57.151866 4968 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.151:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:25:57 crc kubenswrapper[4968]: E0218 15:25:57.228586 4968 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.151:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189560b41862744c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-18 15:25:55.059643468 +0000 UTC m=+254.445088370,LastTimestamp:2026-02-18 15:25:55.059643468 +0000 UTC m=+254.445088370,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 18 15:25:57 crc kubenswrapper[4968]: E0218 15:25:57.942240 4968 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.151:6443: connect: connection refused" interval="6.4s" Feb 18 15:26:01 crc kubenswrapper[4968]: I0218 15:26:01.235417 4968 status_manager.go:851] "Failed to get status for pod" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:26:03 crc kubenswrapper[4968]: I0218 15:26:03.230329 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:26:03 crc kubenswrapper[4968]: I0218 15:26:03.231801 4968 status_manager.go:851] "Failed to get status for pod" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:26:03 crc kubenswrapper[4968]: I0218 15:26:03.250320 4968 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="22c34049-85fe-4d7b-af06-64a247724267" Feb 18 15:26:03 crc kubenswrapper[4968]: I0218 15:26:03.250379 4968 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="22c34049-85fe-4d7b-af06-64a247724267" Feb 18 15:26:03 crc kubenswrapper[4968]: E0218 15:26:03.251710 4968 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:26:03 crc kubenswrapper[4968]: I0218 15:26:03.252571 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:26:03 crc kubenswrapper[4968]: E0218 15:26:03.663575 4968 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-182a392fa9343655994e5dc4860098ef427f0dc5fae92cef95b0cd4f3ba46630.scope\": RecentStats: unable to find data in memory cache]" Feb 18 15:26:04 crc kubenswrapper[4968]: I0218 15:26:04.191229 4968 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="182a392fa9343655994e5dc4860098ef427f0dc5fae92cef95b0cd4f3ba46630" exitCode=0 Feb 18 15:26:04 crc kubenswrapper[4968]: I0218 15:26:04.191274 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"182a392fa9343655994e5dc4860098ef427f0dc5fae92cef95b0cd4f3ba46630"} Feb 18 15:26:04 crc kubenswrapper[4968]: I0218 15:26:04.191300 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"86d09081844b55d2d5109bd0fa713b27e11896a960111f044a37af5e94854a52"} Feb 18 15:26:04 crc kubenswrapper[4968]: I0218 15:26:04.191545 4968 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="22c34049-85fe-4d7b-af06-64a247724267" Feb 18 15:26:04 crc kubenswrapper[4968]: I0218 15:26:04.191557 4968 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="22c34049-85fe-4d7b-af06-64a247724267" Feb 18 15:26:04 crc kubenswrapper[4968]: I0218 15:26:04.192117 4968 status_manager.go:851] "Failed to get status for pod" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" Feb 18 15:26:04 crc kubenswrapper[4968]: E0218 15:26:04.192132 4968 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.151:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:26:04 crc kubenswrapper[4968]: E0218 15:26:04.343663 4968 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.151:6443: connect: connection refused" interval="7s" Feb 18 15:26:05 crc kubenswrapper[4968]: I0218 15:26:05.137575 4968 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 18 15:26:05 crc kubenswrapper[4968]: I0218 15:26:05.137654 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 18 15:26:05 crc kubenswrapper[4968]: I0218 15:26:05.205026 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a38635e21cc3264c91baf027e40c29ce53175c0a7277b749113fb38e1c25d3f9"} Feb 18 15:26:05 crc kubenswrapper[4968]: I0218 15:26:05.205070 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ef4f5aa48947ae64fea0de53d7538fdeedea6eb7478088832fe10d5ea9528d59"} Feb 18 15:26:05 crc kubenswrapper[4968]: I0218 15:26:05.205080 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"debf043c0ea051322c1cbc388df7e49221b3617a5b6a7fd1b1e285c45614f980"} Feb 18 15:26:05 crc kubenswrapper[4968]: I0218 15:26:05.205090 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"244d070ef1971914ccbeed28c5b354bd6103ec7d0ad843f124f998947cc8a8d6"} Feb 18 15:26:05 crc kubenswrapper[4968]: I0218 15:26:05.208347 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 18 15:26:05 crc kubenswrapper[4968]: I0218 15:26:05.208393 4968 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1" exitCode=1 Feb 18 15:26:05 crc kubenswrapper[4968]: I0218 15:26:05.208425 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1"} Feb 18 15:26:05 crc kubenswrapper[4968]: I0218 15:26:05.208860 4968 scope.go:117] "RemoveContainer" containerID="ebe016cc6549499cc87da2351af49562d418fa2d3a1c794bdaa0adb64edc5fa1" Feb 18 15:26:06 crc kubenswrapper[4968]: I0218 15:26:06.224622 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0c6698af4f2001abd15ac9d86afa4e0a427843608368ac18bf82721dd90f77ad"} Feb 18 15:26:06 crc kubenswrapper[4968]: I0218 15:26:06.225000 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:26:06 crc kubenswrapper[4968]: I0218 15:26:06.224837 4968 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="22c34049-85fe-4d7b-af06-64a247724267" Feb 18 15:26:06 crc kubenswrapper[4968]: I0218 15:26:06.225026 4968 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="22c34049-85fe-4d7b-af06-64a247724267" Feb 18 15:26:06 crc kubenswrapper[4968]: I0218 15:26:06.228736 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 18 15:26:06 crc kubenswrapper[4968]: I0218 15:26:06.228882 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e29bcb1a971f0f67248a4996fef0158700c1ae6a4a817b081549fdce9c9e4b53"} Feb 18 15:26:08 crc kubenswrapper[4968]: I0218 15:26:08.253501 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:26:08 crc kubenswrapper[4968]: I0218 15:26:08.253864 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:26:08 crc kubenswrapper[4968]: I0218 15:26:08.259413 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:26:09 crc kubenswrapper[4968]: I0218 15:26:09.875264 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 15:26:09 crc kubenswrapper[4968]: I0218 15:26:09.880698 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 15:26:10 crc kubenswrapper[4968]: I0218 15:26:10.251513 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 15:26:11 crc kubenswrapper[4968]: I0218 15:26:11.242203 4968 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:26:12 crc kubenswrapper[4968]: I0218 15:26:12.261552 4968 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="22c34049-85fe-4d7b-af06-64a247724267" Feb 18 15:26:12 crc kubenswrapper[4968]: I0218 15:26:12.261904 4968 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="22c34049-85fe-4d7b-af06-64a247724267" Feb 18 15:26:12 crc kubenswrapper[4968]: I0218 15:26:12.265976 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:26:12 crc kubenswrapper[4968]: I0218 15:26:12.269312 4968 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="28bf30da-93d2-43cd-9843-c5d8487429e4" Feb 18 15:26:13 crc kubenswrapper[4968]: I0218 15:26:13.270191 4968 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="22c34049-85fe-4d7b-af06-64a247724267" Feb 18 15:26:13 crc kubenswrapper[4968]: I0218 15:26:13.270246 4968 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="22c34049-85fe-4d7b-af06-64a247724267" Feb 18 15:26:13 crc kubenswrapper[4968]: I0218 15:26:13.272091 4968 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="28bf30da-93d2-43cd-9843-c5d8487429e4" Feb 18 15:26:15 crc kubenswrapper[4968]: I0218 15:26:15.143704 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 15:26:21 crc kubenswrapper[4968]: I0218 15:26:21.028882 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 18 15:26:21 crc kubenswrapper[4968]: I0218 15:26:21.221597 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 18 15:26:21 crc kubenswrapper[4968]: I0218 15:26:21.553838 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 18 15:26:22 crc kubenswrapper[4968]: I0218 15:26:22.256251 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 18 15:26:22 crc kubenswrapper[4968]: I0218 15:26:22.320464 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 18 15:26:22 crc kubenswrapper[4968]: I0218 15:26:22.424578 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 18 15:26:22 crc kubenswrapper[4968]: I0218 15:26:22.734916 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 18 15:26:22 crc kubenswrapper[4968]: I0218 15:26:22.763108 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 18 15:26:22 crc kubenswrapper[4968]: I0218 15:26:22.803781 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 18 15:26:22 crc kubenswrapper[4968]: I0218 15:26:22.804969 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 18 15:26:23 crc kubenswrapper[4968]: I0218 15:26:23.100712 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 18 15:26:23 crc kubenswrapper[4968]: I0218 15:26:23.104014 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 18 15:26:23 crc kubenswrapper[4968]: I0218 15:26:23.170551 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 18 15:26:23 crc kubenswrapper[4968]: I0218 15:26:23.197417 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 18 15:26:23 crc kubenswrapper[4968]: I0218 15:26:23.243411 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 18 15:26:23 crc kubenswrapper[4968]: I0218 15:26:23.760300 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 18 15:26:23 crc kubenswrapper[4968]: I0218 15:26:23.894597 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 18 15:26:23 crc kubenswrapper[4968]: I0218 15:26:23.981165 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:23.999962 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.018836 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.111723 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.200897 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.264478 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.484044 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.545712 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.602236 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.614292 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.687212 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.699679 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.736523 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.771517 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.771518 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.833930 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.932924 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 18 15:26:24 crc kubenswrapper[4968]: I0218 15:26:24.969861 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.043741 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.102951 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.178932 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.198768 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.204838 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.259588 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.360522 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.411292 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.573563 4968 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.574825 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.614571 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.643723 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.784255 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.883017 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.884115 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.969665 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 18 15:26:25 crc kubenswrapper[4968]: I0218 15:26:25.986618 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.127041 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.141556 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.147893 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.152601 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.155315 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.161217 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.179146 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.187114 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.198890 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.219835 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.432239 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.440384 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.533492 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.539381 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.610786 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.625114 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.688504 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.767138 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.782729 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 18 15:26:26 crc kubenswrapper[4968]: I0218 15:26:26.994984 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.157945 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.184807 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.203235 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.241665 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.261306 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.282548 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.307539 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.476719 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.650804 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.710774 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.729645 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.793059 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.846225 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.901301 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.931797 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 18 15:26:27 crc kubenswrapper[4968]: I0218 15:26:27.954259 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.020438 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.038274 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.088363 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.099017 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.099150 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.209730 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.238807 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.454414 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.500327 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.688928 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.730014 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.741470 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.761941 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.855824 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.897232 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.925335 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 18 15:26:28 crc kubenswrapper[4968]: I0218 15:26:28.994267 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.055268 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.139262 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.157886 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.172245 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.215612 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.273850 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.293556 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.295117 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.315383 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.363735 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.465899 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.490376 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.515631 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.516451 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.525038 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.547476 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.552843 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.625737 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.646595 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.668409 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.670409 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.719520 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.773174 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.774666 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.823297 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.878626 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 18 15:26:29 crc kubenswrapper[4968]: I0218 15:26:29.928677 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.050990 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.072004 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.163197 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.205326 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.264066 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.275999 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.296033 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.311875 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.316601 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.348893 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.360681 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.459878 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.496823 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.501325 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.515299 4968 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.518225 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.520340 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.570904 4968 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.602471 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.607094 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.620571 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.626921 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.639945 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.680527 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.817763 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.862671 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.884810 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.911546 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.922875 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.927154 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.963606 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.972602 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 18 15:26:30 crc kubenswrapper[4968]: I0218 15:26:30.992420 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.123290 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.141228 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.159601 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.279270 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.296253 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.300958 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.351054 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.352623 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.429068 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.473275 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.488773 4968 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.494640 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.494711 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.500826 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.514614 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.514592595 podStartE2EDuration="20.514592595s" podCreationTimestamp="2026-02-18 15:26:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:26:31.513280138 +0000 UTC m=+290.898725010" watchObservedRunningTime="2026-02-18 15:26:31.514592595 +0000 UTC m=+290.900037457" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.572809 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.608496 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.672691 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.785894 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.819856 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.848821 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.883241 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.908132 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.949283 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.951142 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 18 15:26:31 crc kubenswrapper[4968]: I0218 15:26:31.954473 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.074722 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.153496 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.371557 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.387554 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.598982 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.643258 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.653794 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.728001 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.786475 4968 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.844150 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.856908 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.907105 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.920225 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.983322 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 18 15:26:32 crc kubenswrapper[4968]: I0218 15:26:32.992482 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.004810 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.021358 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.306247 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.498260 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.513920 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.565925 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.567328 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.606257 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.661313 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.688970 4968 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.689216 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://82eaa531149f6c4f6cbe9b4c5816880d95efa5bacf479bdbfe8c55bf3a452b38" gracePeriod=5 Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.690154 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.698266 4968 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.965285 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 18 15:26:33 crc kubenswrapper[4968]: I0218 15:26:33.977250 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.029699 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.135845 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.191858 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.207183 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.267073 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.374541 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.451080 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.470819 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.480494 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.615012 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.625365 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.677891 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.689820 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.730560 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.740271 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.764361 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.766487 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.816257 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 18 15:26:34 crc kubenswrapper[4968]: I0218 15:26:34.879876 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 18 15:26:35 crc kubenswrapper[4968]: I0218 15:26:35.061700 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 18 15:26:35 crc kubenswrapper[4968]: I0218 15:26:35.188985 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 18 15:26:35 crc kubenswrapper[4968]: I0218 15:26:35.245991 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 18 15:26:35 crc kubenswrapper[4968]: I0218 15:26:35.515613 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 18 15:26:35 crc kubenswrapper[4968]: I0218 15:26:35.617141 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 18 15:26:35 crc kubenswrapper[4968]: I0218 15:26:35.737170 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 18 15:26:35 crc kubenswrapper[4968]: I0218 15:26:35.948987 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 18 15:26:36 crc kubenswrapper[4968]: I0218 15:26:36.224997 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 18 15:26:36 crc kubenswrapper[4968]: I0218 15:26:36.292427 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 18 15:26:36 crc kubenswrapper[4968]: I0218 15:26:36.541306 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 18 15:26:36 crc kubenswrapper[4968]: I0218 15:26:36.594378 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 18 15:26:37 crc kubenswrapper[4968]: I0218 15:26:37.012009 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 18 15:26:37 crc kubenswrapper[4968]: I0218 15:26:37.288846 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 18 15:26:37 crc kubenswrapper[4968]: I0218 15:26:37.708466 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 18 15:26:37 crc kubenswrapper[4968]: I0218 15:26:37.978541 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 18 15:26:38 crc kubenswrapper[4968]: I0218 15:26:38.250569 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.178371 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.262122 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.262234 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.303514 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.303556 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.303578 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.303591 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.303631 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.303674 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.303707 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.303707 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.303726 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.304207 4968 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.304224 4968 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.304236 4968 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.304248 4968 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.313548 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.404819 4968 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.413402 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.413462 4968 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="82eaa531149f6c4f6cbe9b4c5816880d95efa5bacf479bdbfe8c55bf3a452b38" exitCode=137 Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.413532 4968 scope.go:117] "RemoveContainer" containerID="82eaa531149f6c4f6cbe9b4c5816880d95efa5bacf479bdbfe8c55bf3a452b38" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.413532 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.438562 4968 scope.go:117] "RemoveContainer" containerID="82eaa531149f6c4f6cbe9b4c5816880d95efa5bacf479bdbfe8c55bf3a452b38" Feb 18 15:26:39 crc kubenswrapper[4968]: E0218 15:26:39.439023 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82eaa531149f6c4f6cbe9b4c5816880d95efa5bacf479bdbfe8c55bf3a452b38\": container with ID starting with 82eaa531149f6c4f6cbe9b4c5816880d95efa5bacf479bdbfe8c55bf3a452b38 not found: ID does not exist" containerID="82eaa531149f6c4f6cbe9b4c5816880d95efa5bacf479bdbfe8c55bf3a452b38" Feb 18 15:26:39 crc kubenswrapper[4968]: I0218 15:26:39.439064 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82eaa531149f6c4f6cbe9b4c5816880d95efa5bacf479bdbfe8c55bf3a452b38"} err="failed to get container status \"82eaa531149f6c4f6cbe9b4c5816880d95efa5bacf479bdbfe8c55bf3a452b38\": rpc error: code = NotFound desc = could not find container \"82eaa531149f6c4f6cbe9b4c5816880d95efa5bacf479bdbfe8c55bf3a452b38\": container with ID starting with 82eaa531149f6c4f6cbe9b4c5816880d95efa5bacf479bdbfe8c55bf3a452b38 not found: ID does not exist" Feb 18 15:26:40 crc kubenswrapper[4968]: I0218 15:26:40.213464 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 18 15:26:41 crc kubenswrapper[4968]: I0218 15:26:41.023369 4968 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 18 15:26:41 crc kubenswrapper[4968]: I0218 15:26:41.239803 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 18 15:26:50 crc kubenswrapper[4968]: I0218 15:26:50.480870 4968 generic.go:334] "Generic (PLEG): container finished" podID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" containerID="a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212" exitCode=0 Feb 18 15:26:50 crc kubenswrapper[4968]: I0218 15:26:50.480987 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" event={"ID":"cc03d1b2-f3f3-4b18-b108-e1a166a6a250","Type":"ContainerDied","Data":"a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212"} Feb 18 15:26:50 crc kubenswrapper[4968]: I0218 15:26:50.482381 4968 scope.go:117] "RemoveContainer" containerID="a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212" Feb 18 15:26:51 crc kubenswrapper[4968]: I0218 15:26:51.486876 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" event={"ID":"cc03d1b2-f3f3-4b18-b108-e1a166a6a250","Type":"ContainerStarted","Data":"d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603"} Feb 18 15:26:51 crc kubenswrapper[4968]: I0218 15:26:51.487253 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:26:51 crc kubenswrapper[4968]: I0218 15:26:51.492889 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.386596 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7b762"] Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.387269 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" podUID="d8e923a2-9b5c-49db-8a02-5eff8ca31033" containerName="controller-manager" containerID="cri-o://3d4acda592e3e185edc66c28d123e85a21d9d5e4ec6ee31c4c9304483298435d" gracePeriod=30 Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.499713 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66"] Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.499952 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" podUID="39580550-6421-4eb7-8727-42c26ad5c3f2" containerName="route-controller-manager" containerID="cri-o://0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1" gracePeriod=30 Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.544004 4968 generic.go:334] "Generic (PLEG): container finished" podID="d8e923a2-9b5c-49db-8a02-5eff8ca31033" containerID="3d4acda592e3e185edc66c28d123e85a21d9d5e4ec6ee31c4c9304483298435d" exitCode=0 Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.544273 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" event={"ID":"d8e923a2-9b5c-49db-8a02-5eff8ca31033","Type":"ContainerDied","Data":"3d4acda592e3e185edc66c28d123e85a21d9d5e4ec6ee31c4c9304483298435d"} Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.711256 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.821431 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-config\") pod \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.821505 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vljcg\" (UniqueName: \"kubernetes.io/projected/d8e923a2-9b5c-49db-8a02-5eff8ca31033-kube-api-access-vljcg\") pod \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.821551 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-proxy-ca-bundles\") pod \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.821572 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-client-ca\") pod \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.821648 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e923a2-9b5c-49db-8a02-5eff8ca31033-serving-cert\") pod \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\" (UID: \"d8e923a2-9b5c-49db-8a02-5eff8ca31033\") " Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.822243 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d8e923a2-9b5c-49db-8a02-5eff8ca31033" (UID: "d8e923a2-9b5c-49db-8a02-5eff8ca31033"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.822248 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-config" (OuterVolumeSpecName: "config") pod "d8e923a2-9b5c-49db-8a02-5eff8ca31033" (UID: "d8e923a2-9b5c-49db-8a02-5eff8ca31033"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.822593 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-client-ca" (OuterVolumeSpecName: "client-ca") pod "d8e923a2-9b5c-49db-8a02-5eff8ca31033" (UID: "d8e923a2-9b5c-49db-8a02-5eff8ca31033"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.827489 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8e923a2-9b5c-49db-8a02-5eff8ca31033-kube-api-access-vljcg" (OuterVolumeSpecName: "kube-api-access-vljcg") pod "d8e923a2-9b5c-49db-8a02-5eff8ca31033" (UID: "d8e923a2-9b5c-49db-8a02-5eff8ca31033"). InnerVolumeSpecName "kube-api-access-vljcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.827532 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8e923a2-9b5c-49db-8a02-5eff8ca31033-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d8e923a2-9b5c-49db-8a02-5eff8ca31033" (UID: "d8e923a2-9b5c-49db-8a02-5eff8ca31033"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.839618 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.922891 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85chc\" (UniqueName: \"kubernetes.io/projected/39580550-6421-4eb7-8727-42c26ad5c3f2-kube-api-access-85chc\") pod \"39580550-6421-4eb7-8727-42c26ad5c3f2\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.922974 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39580550-6421-4eb7-8727-42c26ad5c3f2-config\") pod \"39580550-6421-4eb7-8727-42c26ad5c3f2\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.923026 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39580550-6421-4eb7-8727-42c26ad5c3f2-client-ca\") pod \"39580550-6421-4eb7-8727-42c26ad5c3f2\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.923074 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39580550-6421-4eb7-8727-42c26ad5c3f2-serving-cert\") pod \"39580550-6421-4eb7-8727-42c26ad5c3f2\" (UID: \"39580550-6421-4eb7-8727-42c26ad5c3f2\") " Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.924055 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.924086 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vljcg\" (UniqueName: \"kubernetes.io/projected/d8e923a2-9b5c-49db-8a02-5eff8ca31033-kube-api-access-vljcg\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.924098 4968 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.924109 4968 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8e923a2-9b5c-49db-8a02-5eff8ca31033-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.924096 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39580550-6421-4eb7-8727-42c26ad5c3f2-config" (OuterVolumeSpecName: "config") pod "39580550-6421-4eb7-8727-42c26ad5c3f2" (UID: "39580550-6421-4eb7-8727-42c26ad5c3f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.924120 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e923a2-9b5c-49db-8a02-5eff8ca31033-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.924108 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39580550-6421-4eb7-8727-42c26ad5c3f2-client-ca" (OuterVolumeSpecName: "client-ca") pod "39580550-6421-4eb7-8727-42c26ad5c3f2" (UID: "39580550-6421-4eb7-8727-42c26ad5c3f2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.926636 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39580550-6421-4eb7-8727-42c26ad5c3f2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "39580550-6421-4eb7-8727-42c26ad5c3f2" (UID: "39580550-6421-4eb7-8727-42c26ad5c3f2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:27:01 crc kubenswrapper[4968]: I0218 15:27:01.926733 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39580550-6421-4eb7-8727-42c26ad5c3f2-kube-api-access-85chc" (OuterVolumeSpecName: "kube-api-access-85chc") pod "39580550-6421-4eb7-8727-42c26ad5c3f2" (UID: "39580550-6421-4eb7-8727-42c26ad5c3f2"). InnerVolumeSpecName "kube-api-access-85chc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.024737 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85chc\" (UniqueName: \"kubernetes.io/projected/39580550-6421-4eb7-8727-42c26ad5c3f2-kube-api-access-85chc\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.024803 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39580550-6421-4eb7-8727-42c26ad5c3f2-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.024822 4968 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39580550-6421-4eb7-8727-42c26ad5c3f2-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.024835 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39580550-6421-4eb7-8727-42c26ad5c3f2-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.203301 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6f9486d74b-j6nl9"] Feb 18 15:27:02 crc kubenswrapper[4968]: E0218 15:27:02.203577 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" containerName="installer" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.203595 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" containerName="installer" Feb 18 15:27:02 crc kubenswrapper[4968]: E0218 15:27:02.203605 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.203614 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 18 15:27:02 crc kubenswrapper[4968]: E0218 15:27:02.203629 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8e923a2-9b5c-49db-8a02-5eff8ca31033" containerName="controller-manager" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.203637 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8e923a2-9b5c-49db-8a02-5eff8ca31033" containerName="controller-manager" Feb 18 15:27:02 crc kubenswrapper[4968]: E0218 15:27:02.203649 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39580550-6421-4eb7-8727-42c26ad5c3f2" containerName="route-controller-manager" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.203656 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="39580550-6421-4eb7-8727-42c26ad5c3f2" containerName="route-controller-manager" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.203769 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.203778 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="39580550-6421-4eb7-8727-42c26ad5c3f2" containerName="route-controller-manager" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.203786 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8e923a2-9b5c-49db-8a02-5eff8ca31033" containerName="controller-manager" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.203797 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="3802ffab-07cd-449d-b890-4b25e1571d57" containerName="installer" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.204128 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.218185 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f9486d74b-j6nl9"] Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.226627 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-serving-cert\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.226687 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-proxy-ca-bundles\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.226718 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-client-ca\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.226792 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fbcm\" (UniqueName: \"kubernetes.io/projected/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-kube-api-access-2fbcm\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.226825 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-config\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.303449 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd"] Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.304310 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.316032 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd"] Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.330116 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fbcm\" (UniqueName: \"kubernetes.io/projected/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-kube-api-access-2fbcm\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.330517 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-config\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.330735 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-serving-cert\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.330854 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-proxy-ca-bundles\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.330925 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-client-ca\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.331933 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-client-ca\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.332227 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-config\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.333008 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-proxy-ca-bundles\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.337944 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-serving-cert\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.349293 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fbcm\" (UniqueName: \"kubernetes.io/projected/b8b4c5dd-b08d-4c5d-94f4-eb7611396011-kube-api-access-2fbcm\") pod \"controller-manager-6f9486d74b-j6nl9\" (UID: \"b8b4c5dd-b08d-4c5d-94f4-eb7611396011\") " pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.431847 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d723d001-4fab-422d-b57c-b23694d70b46-client-ca\") pod \"route-controller-manager-59f888f49c-flgtd\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.432209 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6ts2\" (UniqueName: \"kubernetes.io/projected/d723d001-4fab-422d-b57c-b23694d70b46-kube-api-access-r6ts2\") pod \"route-controller-manager-59f888f49c-flgtd\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.432234 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d723d001-4fab-422d-b57c-b23694d70b46-config\") pod \"route-controller-manager-59f888f49c-flgtd\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.432272 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d723d001-4fab-422d-b57c-b23694d70b46-serving-cert\") pod \"route-controller-manager-59f888f49c-flgtd\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.533225 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d723d001-4fab-422d-b57c-b23694d70b46-client-ca\") pod \"route-controller-manager-59f888f49c-flgtd\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.533299 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6ts2\" (UniqueName: \"kubernetes.io/projected/d723d001-4fab-422d-b57c-b23694d70b46-kube-api-access-r6ts2\") pod \"route-controller-manager-59f888f49c-flgtd\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.533329 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d723d001-4fab-422d-b57c-b23694d70b46-config\") pod \"route-controller-manager-59f888f49c-flgtd\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.533360 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d723d001-4fab-422d-b57c-b23694d70b46-serving-cert\") pod \"route-controller-manager-59f888f49c-flgtd\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.534391 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d723d001-4fab-422d-b57c-b23694d70b46-client-ca\") pod \"route-controller-manager-59f888f49c-flgtd\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.534876 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d723d001-4fab-422d-b57c-b23694d70b46-config\") pod \"route-controller-manager-59f888f49c-flgtd\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.538510 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d723d001-4fab-422d-b57c-b23694d70b46-serving-cert\") pod \"route-controller-manager-59f888f49c-flgtd\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.551356 4968 generic.go:334] "Generic (PLEG): container finished" podID="39580550-6421-4eb7-8727-42c26ad5c3f2" containerID="0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1" exitCode=0 Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.551413 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.551414 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" event={"ID":"39580550-6421-4eb7-8727-42c26ad5c3f2","Type":"ContainerDied","Data":"0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1"} Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.551479 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66" event={"ID":"39580550-6421-4eb7-8727-42c26ad5c3f2","Type":"ContainerDied","Data":"3b56b61e95ee58adbf4461ff17d2a859d0ef633b8ac19776bae8bce931f9d84a"} Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.551502 4968 scope.go:117] "RemoveContainer" containerID="0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.555239 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.562301 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" event={"ID":"d8e923a2-9b5c-49db-8a02-5eff8ca31033","Type":"ContainerDied","Data":"5d78c8879c2bd5e1db2b40acf62fe01315492fc86172847957b9d2ae87475c06"} Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.562389 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7b762" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.565318 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6ts2\" (UniqueName: \"kubernetes.io/projected/d723d001-4fab-422d-b57c-b23694d70b46-kube-api-access-r6ts2\") pod \"route-controller-manager-59f888f49c-flgtd\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.572940 4968 scope.go:117] "RemoveContainer" containerID="0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1" Feb 18 15:27:02 crc kubenswrapper[4968]: E0218 15:27:02.574029 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1\": container with ID starting with 0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1 not found: ID does not exist" containerID="0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.574073 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1"} err="failed to get container status \"0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1\": rpc error: code = NotFound desc = could not find container \"0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1\": container with ID starting with 0b611955a42d8f947cb221bd88e9efce71cf03028b6778e78ca6796e43727ff1 not found: ID does not exist" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.574099 4968 scope.go:117] "RemoveContainer" containerID="3d4acda592e3e185edc66c28d123e85a21d9d5e4ec6ee31c4c9304483298435d" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.611512 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7b762"] Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.616240 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7b762"] Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.620117 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.626843 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66"] Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.633407 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f7s66"] Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.740617 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f9486d74b-j6nl9"] Feb 18 15:27:02 crc kubenswrapper[4968]: I0218 15:27:02.823880 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd"] Feb 18 15:27:02 crc kubenswrapper[4968]: W0218 15:27:02.833778 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd723d001_4fab_422d_b57c_b23694d70b46.slice/crio-ea8fbd95a4a916d42fa4e7ce1297513b1b59a4deaf8c67395b195ae3d30c650c WatchSource:0}: Error finding container ea8fbd95a4a916d42fa4e7ce1297513b1b59a4deaf8c67395b195ae3d30c650c: Status 404 returned error can't find the container with id ea8fbd95a4a916d42fa4e7ce1297513b1b59a4deaf8c67395b195ae3d30c650c Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.028555 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd"] Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.237435 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39580550-6421-4eb7-8727-42c26ad5c3f2" path="/var/lib/kubelet/pods/39580550-6421-4eb7-8727-42c26ad5c3f2/volumes" Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.238021 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8e923a2-9b5c-49db-8a02-5eff8ca31033" path="/var/lib/kubelet/pods/d8e923a2-9b5c-49db-8a02-5eff8ca31033/volumes" Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.569466 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" event={"ID":"d723d001-4fab-422d-b57c-b23694d70b46","Type":"ContainerStarted","Data":"832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565"} Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.569815 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" event={"ID":"d723d001-4fab-422d-b57c-b23694d70b46","Type":"ContainerStarted","Data":"ea8fbd95a4a916d42fa4e7ce1297513b1b59a4deaf8c67395b195ae3d30c650c"} Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.570163 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.572822 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" event={"ID":"b8b4c5dd-b08d-4c5d-94f4-eb7611396011","Type":"ContainerStarted","Data":"a91a955d959cd61654eae05fac0e58c13b5f07e1703b5579ea6bdf24f41bfeb5"} Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.572850 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" event={"ID":"b8b4c5dd-b08d-4c5d-94f4-eb7611396011","Type":"ContainerStarted","Data":"297e4daaa18021355c91c93aa7e2c95e016b5adaf409bef272c5f9f114cdaefb"} Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.573013 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.576273 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.578638 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.584469 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" podStartSLOduration=1.5844529170000001 podStartE2EDuration="1.584452917s" podCreationTimestamp="2026-02-18 15:27:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:27:03.582930274 +0000 UTC m=+322.968375136" watchObservedRunningTime="2026-02-18 15:27:03.584452917 +0000 UTC m=+322.969897779" Feb 18 15:27:03 crc kubenswrapper[4968]: I0218 15:27:03.597845 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6f9486d74b-j6nl9" podStartSLOduration=1.5978254170000001 podStartE2EDuration="1.597825417s" podCreationTimestamp="2026-02-18 15:27:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:27:03.593623238 +0000 UTC m=+322.979068100" watchObservedRunningTime="2026-02-18 15:27:03.597825417 +0000 UTC m=+322.983270279" Feb 18 15:27:04 crc kubenswrapper[4968]: I0218 15:27:04.578427 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" podUID="d723d001-4fab-422d-b57c-b23694d70b46" containerName="route-controller-manager" containerID="cri-o://832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565" gracePeriod=30 Feb 18 15:27:04 crc kubenswrapper[4968]: I0218 15:27:04.955875 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:04 crc kubenswrapper[4968]: I0218 15:27:04.978972 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c"] Feb 18 15:27:04 crc kubenswrapper[4968]: E0218 15:27:04.979184 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d723d001-4fab-422d-b57c-b23694d70b46" containerName="route-controller-manager" Feb 18 15:27:04 crc kubenswrapper[4968]: I0218 15:27:04.979196 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="d723d001-4fab-422d-b57c-b23694d70b46" containerName="route-controller-manager" Feb 18 15:27:04 crc kubenswrapper[4968]: I0218 15:27:04.979281 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="d723d001-4fab-422d-b57c-b23694d70b46" containerName="route-controller-manager" Feb 18 15:27:04 crc kubenswrapper[4968]: I0218 15:27:04.979618 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.001462 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c"] Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.064513 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6ts2\" (UniqueName: \"kubernetes.io/projected/d723d001-4fab-422d-b57c-b23694d70b46-kube-api-access-r6ts2\") pod \"d723d001-4fab-422d-b57c-b23694d70b46\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.064658 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d723d001-4fab-422d-b57c-b23694d70b46-serving-cert\") pod \"d723d001-4fab-422d-b57c-b23694d70b46\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.064771 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d723d001-4fab-422d-b57c-b23694d70b46-client-ca\") pod \"d723d001-4fab-422d-b57c-b23694d70b46\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.064861 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d723d001-4fab-422d-b57c-b23694d70b46-config\") pod \"d723d001-4fab-422d-b57c-b23694d70b46\" (UID: \"d723d001-4fab-422d-b57c-b23694d70b46\") " Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.065650 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d723d001-4fab-422d-b57c-b23694d70b46-client-ca" (OuterVolumeSpecName: "client-ca") pod "d723d001-4fab-422d-b57c-b23694d70b46" (UID: "d723d001-4fab-422d-b57c-b23694d70b46"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.065675 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d723d001-4fab-422d-b57c-b23694d70b46-config" (OuterVolumeSpecName: "config") pod "d723d001-4fab-422d-b57c-b23694d70b46" (UID: "d723d001-4fab-422d-b57c-b23694d70b46"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.071986 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d723d001-4fab-422d-b57c-b23694d70b46-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d723d001-4fab-422d-b57c-b23694d70b46" (UID: "d723d001-4fab-422d-b57c-b23694d70b46"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.072038 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d723d001-4fab-422d-b57c-b23694d70b46-kube-api-access-r6ts2" (OuterVolumeSpecName: "kube-api-access-r6ts2") pod "d723d001-4fab-422d-b57c-b23694d70b46" (UID: "d723d001-4fab-422d-b57c-b23694d70b46"). InnerVolumeSpecName "kube-api-access-r6ts2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.166880 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7c0f8d9b-70bc-428f-934a-494c834be4c4-client-ca\") pod \"route-controller-manager-6cd979f5b6-jhx7c\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.166950 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9zsp\" (UniqueName: \"kubernetes.io/projected/7c0f8d9b-70bc-428f-934a-494c834be4c4-kube-api-access-x9zsp\") pod \"route-controller-manager-6cd979f5b6-jhx7c\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.167064 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c0f8d9b-70bc-428f-934a-494c834be4c4-serving-cert\") pod \"route-controller-manager-6cd979f5b6-jhx7c\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.167121 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c0f8d9b-70bc-428f-934a-494c834be4c4-config\") pod \"route-controller-manager-6cd979f5b6-jhx7c\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.167172 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d723d001-4fab-422d-b57c-b23694d70b46-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.167190 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6ts2\" (UniqueName: \"kubernetes.io/projected/d723d001-4fab-422d-b57c-b23694d70b46-kube-api-access-r6ts2\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.167207 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d723d001-4fab-422d-b57c-b23694d70b46-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.167222 4968 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d723d001-4fab-422d-b57c-b23694d70b46-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.268350 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7c0f8d9b-70bc-428f-934a-494c834be4c4-client-ca\") pod \"route-controller-manager-6cd979f5b6-jhx7c\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.268418 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9zsp\" (UniqueName: \"kubernetes.io/projected/7c0f8d9b-70bc-428f-934a-494c834be4c4-kube-api-access-x9zsp\") pod \"route-controller-manager-6cd979f5b6-jhx7c\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.269325 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c0f8d9b-70bc-428f-934a-494c834be4c4-serving-cert\") pod \"route-controller-manager-6cd979f5b6-jhx7c\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.269361 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c0f8d9b-70bc-428f-934a-494c834be4c4-config\") pod \"route-controller-manager-6cd979f5b6-jhx7c\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.270651 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c0f8d9b-70bc-428f-934a-494c834be4c4-config\") pod \"route-controller-manager-6cd979f5b6-jhx7c\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.270988 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7c0f8d9b-70bc-428f-934a-494c834be4c4-client-ca\") pod \"route-controller-manager-6cd979f5b6-jhx7c\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.274052 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c0f8d9b-70bc-428f-934a-494c834be4c4-serving-cert\") pod \"route-controller-manager-6cd979f5b6-jhx7c\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.287692 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9zsp\" (UniqueName: \"kubernetes.io/projected/7c0f8d9b-70bc-428f-934a-494c834be4c4-kube-api-access-x9zsp\") pod \"route-controller-manager-6cd979f5b6-jhx7c\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.304658 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.584049 4968 generic.go:334] "Generic (PLEG): container finished" podID="d723d001-4fab-422d-b57c-b23694d70b46" containerID="832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565" exitCode=0 Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.584807 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.584920 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" event={"ID":"d723d001-4fab-422d-b57c-b23694d70b46","Type":"ContainerDied","Data":"832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565"} Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.584980 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd" event={"ID":"d723d001-4fab-422d-b57c-b23694d70b46","Type":"ContainerDied","Data":"ea8fbd95a4a916d42fa4e7ce1297513b1b59a4deaf8c67395b195ae3d30c650c"} Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.584997 4968 scope.go:117] "RemoveContainer" containerID="832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.600197 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd"] Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.604127 4968 scope.go:117] "RemoveContainer" containerID="832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565" Feb 18 15:27:05 crc kubenswrapper[4968]: E0218 15:27:05.604539 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565\": container with ID starting with 832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565 not found: ID does not exist" containerID="832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.604570 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565"} err="failed to get container status \"832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565\": rpc error: code = NotFound desc = could not find container \"832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565\": container with ID starting with 832e8609e0035492cdd822c982f4f95bd0ce984732a85f3f49ba98d71212b565 not found: ID does not exist" Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.607394 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f888f49c-flgtd"] Feb 18 15:27:05 crc kubenswrapper[4968]: I0218 15:27:05.741266 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c"] Feb 18 15:27:05 crc kubenswrapper[4968]: W0218 15:27:05.747616 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c0f8d9b_70bc_428f_934a_494c834be4c4.slice/crio-a88f6766789cec234a89f9145e6439ccfcf8f30f921b082eeb5c1da32a69fe44 WatchSource:0}: Error finding container a88f6766789cec234a89f9145e6439ccfcf8f30f921b082eeb5c1da32a69fe44: Status 404 returned error can't find the container with id a88f6766789cec234a89f9145e6439ccfcf8f30f921b082eeb5c1da32a69fe44 Feb 18 15:27:06 crc kubenswrapper[4968]: I0218 15:27:06.590722 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" event={"ID":"7c0f8d9b-70bc-428f-934a-494c834be4c4","Type":"ContainerStarted","Data":"3afd90258bf18678b54ba2c604ad874850b51818d6444813ea9ff4377f74d2d8"} Feb 18 15:27:06 crc kubenswrapper[4968]: I0218 15:27:06.591067 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:06 crc kubenswrapper[4968]: I0218 15:27:06.591081 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" event={"ID":"7c0f8d9b-70bc-428f-934a-494c834be4c4","Type":"ContainerStarted","Data":"a88f6766789cec234a89f9145e6439ccfcf8f30f921b082eeb5c1da32a69fe44"} Feb 18 15:27:06 crc kubenswrapper[4968]: I0218 15:27:06.595843 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:06 crc kubenswrapper[4968]: I0218 15:27:06.609086 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" podStartSLOduration=3.609066264 podStartE2EDuration="3.609066264s" podCreationTimestamp="2026-02-18 15:27:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:27:06.608525129 +0000 UTC m=+325.993969991" watchObservedRunningTime="2026-02-18 15:27:06.609066264 +0000 UTC m=+325.994511126" Feb 18 15:27:07 crc kubenswrapper[4968]: I0218 15:27:07.237194 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d723d001-4fab-422d-b57c-b23694d70b46" path="/var/lib/kubelet/pods/d723d001-4fab-422d-b57c-b23694d70b46/volumes" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.486116 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-s4wpl"] Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.487542 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.497808 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-s4wpl"] Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.619829 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-registry-tls\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.619905 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-registry-certificates\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.619947 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-trusted-ca\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.619991 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-ca-trust-extracted\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.620031 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfv5k\" (UniqueName: \"kubernetes.io/projected/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-kube-api-access-cfv5k\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.620105 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-installation-pull-secrets\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.620158 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.620223 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-bound-sa-token\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.643350 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.721711 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-bound-sa-token\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.721811 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-registry-tls\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.721842 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-registry-certificates\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.721872 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-trusted-ca\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.721909 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-ca-trust-extracted\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.721929 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfv5k\" (UniqueName: \"kubernetes.io/projected/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-kube-api-access-cfv5k\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.721948 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-installation-pull-secrets\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.722852 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-ca-trust-extracted\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.723362 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-trusted-ca\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.724222 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-registry-certificates\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.729627 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-installation-pull-secrets\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.730299 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-registry-tls\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.744668 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-bound-sa-token\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.745217 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfv5k\" (UniqueName: \"kubernetes.io/projected/ddc6f1fe-3b21-49d3-9afd-14c54b501b04-kube-api-access-cfv5k\") pod \"image-registry-66df7c8f76-s4wpl\" (UID: \"ddc6f1fe-3b21-49d3-9afd-14c54b501b04\") " pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:14 crc kubenswrapper[4968]: I0218 15:27:14.810325 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:15 crc kubenswrapper[4968]: I0218 15:27:15.224400 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-s4wpl"] Feb 18 15:27:15 crc kubenswrapper[4968]: W0218 15:27:15.232067 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddc6f1fe_3b21_49d3_9afd_14c54b501b04.slice/crio-228a444e9813fb81f9012d003bd99e11bdbfeed59facebb2e6f754208f8c9b2a WatchSource:0}: Error finding container 228a444e9813fb81f9012d003bd99e11bdbfeed59facebb2e6f754208f8c9b2a: Status 404 returned error can't find the container with id 228a444e9813fb81f9012d003bd99e11bdbfeed59facebb2e6f754208f8c9b2a Feb 18 15:27:15 crc kubenswrapper[4968]: I0218 15:27:15.647619 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" event={"ID":"ddc6f1fe-3b21-49d3-9afd-14c54b501b04","Type":"ContainerStarted","Data":"8aa720c7a36bd1d60d1f2d9bd2a8261c8d7b6b4fe27603ee34168a994de57817"} Feb 18 15:27:15 crc kubenswrapper[4968]: I0218 15:27:15.647952 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" event={"ID":"ddc6f1fe-3b21-49d3-9afd-14c54b501b04","Type":"ContainerStarted","Data":"228a444e9813fb81f9012d003bd99e11bdbfeed59facebb2e6f754208f8c9b2a"} Feb 18 15:27:15 crc kubenswrapper[4968]: I0218 15:27:15.648087 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:15 crc kubenswrapper[4968]: I0218 15:27:15.671552 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" podStartSLOduration=1.671537128 podStartE2EDuration="1.671537128s" podCreationTimestamp="2026-02-18 15:27:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:27:15.669882632 +0000 UTC m=+335.055327504" watchObservedRunningTime="2026-02-18 15:27:15.671537128 +0000 UTC m=+335.056981990" Feb 18 15:27:21 crc kubenswrapper[4968]: I0218 15:27:21.440152 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c"] Feb 18 15:27:21 crc kubenswrapper[4968]: I0218 15:27:21.440698 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" podUID="7c0f8d9b-70bc-428f-934a-494c834be4c4" containerName="route-controller-manager" containerID="cri-o://3afd90258bf18678b54ba2c604ad874850b51818d6444813ea9ff4377f74d2d8" gracePeriod=30 Feb 18 15:27:21 crc kubenswrapper[4968]: I0218 15:27:21.682842 4968 generic.go:334] "Generic (PLEG): container finished" podID="7c0f8d9b-70bc-428f-934a-494c834be4c4" containerID="3afd90258bf18678b54ba2c604ad874850b51818d6444813ea9ff4377f74d2d8" exitCode=0 Feb 18 15:27:21 crc kubenswrapper[4968]: I0218 15:27:21.682927 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" event={"ID":"7c0f8d9b-70bc-428f-934a-494c834be4c4","Type":"ContainerDied","Data":"3afd90258bf18678b54ba2c604ad874850b51818d6444813ea9ff4377f74d2d8"} Feb 18 15:27:21 crc kubenswrapper[4968]: I0218 15:27:21.862423 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.021800 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c0f8d9b-70bc-428f-934a-494c834be4c4-config\") pod \"7c0f8d9b-70bc-428f-934a-494c834be4c4\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.021873 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7c0f8d9b-70bc-428f-934a-494c834be4c4-client-ca\") pod \"7c0f8d9b-70bc-428f-934a-494c834be4c4\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.021906 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9zsp\" (UniqueName: \"kubernetes.io/projected/7c0f8d9b-70bc-428f-934a-494c834be4c4-kube-api-access-x9zsp\") pod \"7c0f8d9b-70bc-428f-934a-494c834be4c4\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.021956 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c0f8d9b-70bc-428f-934a-494c834be4c4-serving-cert\") pod \"7c0f8d9b-70bc-428f-934a-494c834be4c4\" (UID: \"7c0f8d9b-70bc-428f-934a-494c834be4c4\") " Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.022839 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c0f8d9b-70bc-428f-934a-494c834be4c4-config" (OuterVolumeSpecName: "config") pod "7c0f8d9b-70bc-428f-934a-494c834be4c4" (UID: "7c0f8d9b-70bc-428f-934a-494c834be4c4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.023511 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c0f8d9b-70bc-428f-934a-494c834be4c4-client-ca" (OuterVolumeSpecName: "client-ca") pod "7c0f8d9b-70bc-428f-934a-494c834be4c4" (UID: "7c0f8d9b-70bc-428f-934a-494c834be4c4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.032225 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c0f8d9b-70bc-428f-934a-494c834be4c4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7c0f8d9b-70bc-428f-934a-494c834be4c4" (UID: "7c0f8d9b-70bc-428f-934a-494c834be4c4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.040269 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c0f8d9b-70bc-428f-934a-494c834be4c4-kube-api-access-x9zsp" (OuterVolumeSpecName: "kube-api-access-x9zsp") pod "7c0f8d9b-70bc-428f-934a-494c834be4c4" (UID: "7c0f8d9b-70bc-428f-934a-494c834be4c4"). InnerVolumeSpecName "kube-api-access-x9zsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.123078 4968 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c0f8d9b-70bc-428f-934a-494c834be4c4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.123112 4968 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c0f8d9b-70bc-428f-934a-494c834be4c4-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.123129 4968 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7c0f8d9b-70bc-428f-934a-494c834be4c4-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.123142 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9zsp\" (UniqueName: \"kubernetes.io/projected/7c0f8d9b-70bc-428f-934a-494c834be4c4-kube-api-access-x9zsp\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.690935 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" event={"ID":"7c0f8d9b-70bc-428f-934a-494c834be4c4","Type":"ContainerDied","Data":"a88f6766789cec234a89f9145e6439ccfcf8f30f921b082eeb5c1da32a69fe44"} Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.691276 4968 scope.go:117] "RemoveContainer" containerID="3afd90258bf18678b54ba2c604ad874850b51818d6444813ea9ff4377f74d2d8" Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.691043 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c" Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.724588 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c"] Feb 18 15:27:22 crc kubenswrapper[4968]: I0218 15:27:22.728732 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd979f5b6-jhx7c"] Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.237441 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c0f8d9b-70bc-428f-934a-494c834be4c4" path="/var/lib/kubelet/pods/7c0f8d9b-70bc-428f-934a-494c834be4c4/volumes" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.437895 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk"] Feb 18 15:27:23 crc kubenswrapper[4968]: E0218 15:27:23.438347 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0f8d9b-70bc-428f-934a-494c834be4c4" containerName="route-controller-manager" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.438377 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0f8d9b-70bc-428f-934a-494c834be4c4" containerName="route-controller-manager" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.438587 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c0f8d9b-70bc-428f-934a-494c834be4c4" containerName="route-controller-manager" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.439295 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.441996 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.442040 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.442416 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.443659 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.447496 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.447736 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.450353 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk"] Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.540663 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75mzs\" (UniqueName: \"kubernetes.io/projected/e2a41fc6-6472-4d8d-914a-1f83bea800a9-kube-api-access-75mzs\") pod \"route-controller-manager-59f888f49c-7mjjk\" (UID: \"e2a41fc6-6472-4d8d-914a-1f83bea800a9\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.540809 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2a41fc6-6472-4d8d-914a-1f83bea800a9-serving-cert\") pod \"route-controller-manager-59f888f49c-7mjjk\" (UID: \"e2a41fc6-6472-4d8d-914a-1f83bea800a9\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.540871 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2a41fc6-6472-4d8d-914a-1f83bea800a9-config\") pod \"route-controller-manager-59f888f49c-7mjjk\" (UID: \"e2a41fc6-6472-4d8d-914a-1f83bea800a9\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.540917 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2a41fc6-6472-4d8d-914a-1f83bea800a9-client-ca\") pod \"route-controller-manager-59f888f49c-7mjjk\" (UID: \"e2a41fc6-6472-4d8d-914a-1f83bea800a9\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.642056 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2a41fc6-6472-4d8d-914a-1f83bea800a9-serving-cert\") pod \"route-controller-manager-59f888f49c-7mjjk\" (UID: \"e2a41fc6-6472-4d8d-914a-1f83bea800a9\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.642160 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2a41fc6-6472-4d8d-914a-1f83bea800a9-config\") pod \"route-controller-manager-59f888f49c-7mjjk\" (UID: \"e2a41fc6-6472-4d8d-914a-1f83bea800a9\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.642224 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2a41fc6-6472-4d8d-914a-1f83bea800a9-client-ca\") pod \"route-controller-manager-59f888f49c-7mjjk\" (UID: \"e2a41fc6-6472-4d8d-914a-1f83bea800a9\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.642315 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75mzs\" (UniqueName: \"kubernetes.io/projected/e2a41fc6-6472-4d8d-914a-1f83bea800a9-kube-api-access-75mzs\") pod \"route-controller-manager-59f888f49c-7mjjk\" (UID: \"e2a41fc6-6472-4d8d-914a-1f83bea800a9\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.644539 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2a41fc6-6472-4d8d-914a-1f83bea800a9-client-ca\") pod \"route-controller-manager-59f888f49c-7mjjk\" (UID: \"e2a41fc6-6472-4d8d-914a-1f83bea800a9\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.645105 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2a41fc6-6472-4d8d-914a-1f83bea800a9-config\") pod \"route-controller-manager-59f888f49c-7mjjk\" (UID: \"e2a41fc6-6472-4d8d-914a-1f83bea800a9\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.649941 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2a41fc6-6472-4d8d-914a-1f83bea800a9-serving-cert\") pod \"route-controller-manager-59f888f49c-7mjjk\" (UID: \"e2a41fc6-6472-4d8d-914a-1f83bea800a9\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.665110 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75mzs\" (UniqueName: \"kubernetes.io/projected/e2a41fc6-6472-4d8d-914a-1f83bea800a9-kube-api-access-75mzs\") pod \"route-controller-manager-59f888f49c-7mjjk\" (UID: \"e2a41fc6-6472-4d8d-914a-1f83bea800a9\") " pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:23 crc kubenswrapper[4968]: I0218 15:27:23.772449 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:24 crc kubenswrapper[4968]: I0218 15:27:24.293665 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk"] Feb 18 15:27:24 crc kubenswrapper[4968]: W0218 15:27:24.299671 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2a41fc6_6472_4d8d_914a_1f83bea800a9.slice/crio-3eb5c6b95892c9a3e2d6cbf7c15fb2b03cdcc474307a5b6fa6e0e8045b14fd89 WatchSource:0}: Error finding container 3eb5c6b95892c9a3e2d6cbf7c15fb2b03cdcc474307a5b6fa6e0e8045b14fd89: Status 404 returned error can't find the container with id 3eb5c6b95892c9a3e2d6cbf7c15fb2b03cdcc474307a5b6fa6e0e8045b14fd89 Feb 18 15:27:24 crc kubenswrapper[4968]: I0218 15:27:24.706543 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" event={"ID":"e2a41fc6-6472-4d8d-914a-1f83bea800a9","Type":"ContainerStarted","Data":"94a56464bc5c1b58332cac68e45c16b1c3bd8619c9f0f366ed9a338a4e235e2d"} Feb 18 15:27:24 crc kubenswrapper[4968]: I0218 15:27:24.706835 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" event={"ID":"e2a41fc6-6472-4d8d-914a-1f83bea800a9","Type":"ContainerStarted","Data":"3eb5c6b95892c9a3e2d6cbf7c15fb2b03cdcc474307a5b6fa6e0e8045b14fd89"} Feb 18 15:27:24 crc kubenswrapper[4968]: I0218 15:27:24.706851 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:24 crc kubenswrapper[4968]: I0218 15:27:24.729564 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" podStartSLOduration=3.729543075 podStartE2EDuration="3.729543075s" podCreationTimestamp="2026-02-18 15:27:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:27:24.72257895 +0000 UTC m=+344.108023812" watchObservedRunningTime="2026-02-18 15:27:24.729543075 +0000 UTC m=+344.114987937" Feb 18 15:27:25 crc kubenswrapper[4968]: I0218 15:27:25.020714 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59f888f49c-7mjjk" Feb 18 15:27:34 crc kubenswrapper[4968]: I0218 15:27:34.816439 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-s4wpl" Feb 18 15:27:34 crc kubenswrapper[4968]: I0218 15:27:34.876482 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lcx97"] Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.687834 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4r24p"] Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.688606 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4r24p" podUID="deeab883-52a5-49df-9b71-e3927cf0f2ea" containerName="registry-server" containerID="cri-o://53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455" gracePeriod=30 Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.695884 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hkb2r"] Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.696325 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hkb2r" podUID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" containerName="registry-server" containerID="cri-o://f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a" gracePeriod=30 Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.712150 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dnmb8"] Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.712494 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" podUID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" containerName="marketplace-operator" containerID="cri-o://d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603" gracePeriod=30 Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.725057 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f4xk"] Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.725615 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5f4xk" podUID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" containerName="registry-server" containerID="cri-o://86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e" gracePeriod=30 Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.744953 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-242v6"] Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.745782 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.751361 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mrbbc"] Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.752921 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mrbbc" podUID="26f10928-44af-4829-b9f7-1fe382373a48" containerName="registry-server" containerID="cri-o://8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547" gracePeriod=30 Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.799912 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-242v6"] Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.873131 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2gnk\" (UniqueName: \"kubernetes.io/projected/e2eb1519-86a1-496a-ace0-ee997d43287c-kube-api-access-h2gnk\") pod \"marketplace-operator-79b997595-242v6\" (UID: \"e2eb1519-86a1-496a-ace0-ee997d43287c\") " pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.873201 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e2eb1519-86a1-496a-ace0-ee997d43287c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-242v6\" (UID: \"e2eb1519-86a1-496a-ace0-ee997d43287c\") " pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.873238 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2eb1519-86a1-496a-ace0-ee997d43287c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-242v6\" (UID: \"e2eb1519-86a1-496a-ace0-ee997d43287c\") " pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.984085 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2gnk\" (UniqueName: \"kubernetes.io/projected/e2eb1519-86a1-496a-ace0-ee997d43287c-kube-api-access-h2gnk\") pod \"marketplace-operator-79b997595-242v6\" (UID: \"e2eb1519-86a1-496a-ace0-ee997d43287c\") " pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.984255 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e2eb1519-86a1-496a-ace0-ee997d43287c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-242v6\" (UID: \"e2eb1519-86a1-496a-ace0-ee997d43287c\") " pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.984349 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2eb1519-86a1-496a-ace0-ee997d43287c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-242v6\" (UID: \"e2eb1519-86a1-496a-ace0-ee997d43287c\") " pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.986223 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2eb1519-86a1-496a-ace0-ee997d43287c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-242v6\" (UID: \"e2eb1519-86a1-496a-ace0-ee997d43287c\") " pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:37 crc kubenswrapper[4968]: I0218 15:27:37.990815 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e2eb1519-86a1-496a-ace0-ee997d43287c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-242v6\" (UID: \"e2eb1519-86a1-496a-ace0-ee997d43287c\") " pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.002380 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2gnk\" (UniqueName: \"kubernetes.io/projected/e2eb1519-86a1-496a-ace0-ee997d43287c-kube-api-access-h2gnk\") pod \"marketplace-operator-79b997595-242v6\" (UID: \"e2eb1519-86a1-496a-ace0-ee997d43287c\") " pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.159149 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.180408 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.292590 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.296209 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deeab883-52a5-49df-9b71-e3927cf0f2ea-utilities\") pod \"deeab883-52a5-49df-9b71-e3927cf0f2ea\" (UID: \"deeab883-52a5-49df-9b71-e3927cf0f2ea\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.296308 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrpnd\" (UniqueName: \"kubernetes.io/projected/deeab883-52a5-49df-9b71-e3927cf0f2ea-kube-api-access-hrpnd\") pod \"deeab883-52a5-49df-9b71-e3927cf0f2ea\" (UID: \"deeab883-52a5-49df-9b71-e3927cf0f2ea\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.296380 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deeab883-52a5-49df-9b71-e3927cf0f2ea-catalog-content\") pod \"deeab883-52a5-49df-9b71-e3927cf0f2ea\" (UID: \"deeab883-52a5-49df-9b71-e3927cf0f2ea\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.298025 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deeab883-52a5-49df-9b71-e3927cf0f2ea-utilities" (OuterVolumeSpecName: "utilities") pod "deeab883-52a5-49df-9b71-e3927cf0f2ea" (UID: "deeab883-52a5-49df-9b71-e3927cf0f2ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.302854 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deeab883-52a5-49df-9b71-e3927cf0f2ea-kube-api-access-hrpnd" (OuterVolumeSpecName: "kube-api-access-hrpnd") pod "deeab883-52a5-49df-9b71-e3927cf0f2ea" (UID: "deeab883-52a5-49df-9b71-e3927cf0f2ea"). InnerVolumeSpecName "kube-api-access-hrpnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.327206 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.334556 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.363613 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.372839 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deeab883-52a5-49df-9b71-e3927cf0f2ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "deeab883-52a5-49df-9b71-e3927cf0f2ea" (UID: "deeab883-52a5-49df-9b71-e3927cf0f2ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.400920 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfl8f\" (UniqueName: \"kubernetes.io/projected/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-kube-api-access-sfl8f\") pod \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\" (UID: \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.400967 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-utilities\") pod \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\" (UID: \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.401000 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83659e7b-5ef2-4128-9a9e-4cbe0e332654-catalog-content\") pod \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\" (UID: \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.401044 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrrhc\" (UniqueName: \"kubernetes.io/projected/83659e7b-5ef2-4128-9a9e-4cbe0e332654-kube-api-access-lrrhc\") pod \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\" (UID: \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.401079 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83659e7b-5ef2-4128-9a9e-4cbe0e332654-utilities\") pod \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\" (UID: \"83659e7b-5ef2-4128-9a9e-4cbe0e332654\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.401131 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-catalog-content\") pod \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\" (UID: \"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.401399 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deeab883-52a5-49df-9b71-e3927cf0f2ea-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.401411 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrpnd\" (UniqueName: \"kubernetes.io/projected/deeab883-52a5-49df-9b71-e3927cf0f2ea-kube-api-access-hrpnd\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.401421 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deeab883-52a5-49df-9b71-e3927cf0f2ea-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.416665 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83659e7b-5ef2-4128-9a9e-4cbe0e332654-utilities" (OuterVolumeSpecName: "utilities") pod "83659e7b-5ef2-4128-9a9e-4cbe0e332654" (UID: "83659e7b-5ef2-4128-9a9e-4cbe0e332654"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.417528 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-utilities" (OuterVolumeSpecName: "utilities") pod "48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" (UID: "48c11632-e38d-4cd3-8ac6-9ef42d55b0e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.429362 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-kube-api-access-sfl8f" (OuterVolumeSpecName: "kube-api-access-sfl8f") pod "48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" (UID: "48c11632-e38d-4cd3-8ac6-9ef42d55b0e5"). InnerVolumeSpecName "kube-api-access-sfl8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.438066 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83659e7b-5ef2-4128-9a9e-4cbe0e332654-kube-api-access-lrrhc" (OuterVolumeSpecName: "kube-api-access-lrrhc") pod "83659e7b-5ef2-4128-9a9e-4cbe0e332654" (UID: "83659e7b-5ef2-4128-9a9e-4cbe0e332654"). InnerVolumeSpecName "kube-api-access-lrrhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.450728 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" (UID: "48c11632-e38d-4cd3-8ac6-9ef42d55b0e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.502659 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f10928-44af-4829-b9f7-1fe382373a48-catalog-content\") pod \"26f10928-44af-4829-b9f7-1fe382373a48\" (UID: \"26f10928-44af-4829-b9f7-1fe382373a48\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.502805 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-marketplace-operator-metrics\") pod \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\" (UID: \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.502898 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtcfw\" (UniqueName: \"kubernetes.io/projected/26f10928-44af-4829-b9f7-1fe382373a48-kube-api-access-jtcfw\") pod \"26f10928-44af-4829-b9f7-1fe382373a48\" (UID: \"26f10928-44af-4829-b9f7-1fe382373a48\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.502955 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f10928-44af-4829-b9f7-1fe382373a48-utilities\") pod \"26f10928-44af-4829-b9f7-1fe382373a48\" (UID: \"26f10928-44af-4829-b9f7-1fe382373a48\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.503002 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w7n5\" (UniqueName: \"kubernetes.io/projected/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-kube-api-access-9w7n5\") pod \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\" (UID: \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.503042 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-marketplace-trusted-ca\") pod \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\" (UID: \"cc03d1b2-f3f3-4b18-b108-e1a166a6a250\") " Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.505629 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26f10928-44af-4829-b9f7-1fe382373a48-utilities" (OuterVolumeSpecName: "utilities") pod "26f10928-44af-4829-b9f7-1fe382373a48" (UID: "26f10928-44af-4829-b9f7-1fe382373a48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.505831 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.505951 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfl8f\" (UniqueName: \"kubernetes.io/projected/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-kube-api-access-sfl8f\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.505968 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.505982 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrrhc\" (UniqueName: \"kubernetes.io/projected/83659e7b-5ef2-4128-9a9e-4cbe0e332654-kube-api-access-lrrhc\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.505992 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83659e7b-5ef2-4128-9a9e-4cbe0e332654-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.506348 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "cc03d1b2-f3f3-4b18-b108-e1a166a6a250" (UID: "cc03d1b2-f3f3-4b18-b108-e1a166a6a250"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.508992 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f10928-44af-4829-b9f7-1fe382373a48-kube-api-access-jtcfw" (OuterVolumeSpecName: "kube-api-access-jtcfw") pod "26f10928-44af-4829-b9f7-1fe382373a48" (UID: "26f10928-44af-4829-b9f7-1fe382373a48"). InnerVolumeSpecName "kube-api-access-jtcfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.509439 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-kube-api-access-9w7n5" (OuterVolumeSpecName: "kube-api-access-9w7n5") pod "cc03d1b2-f3f3-4b18-b108-e1a166a6a250" (UID: "cc03d1b2-f3f3-4b18-b108-e1a166a6a250"). InnerVolumeSpecName "kube-api-access-9w7n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.512177 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "cc03d1b2-f3f3-4b18-b108-e1a166a6a250" (UID: "cc03d1b2-f3f3-4b18-b108-e1a166a6a250"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.517702 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83659e7b-5ef2-4128-9a9e-4cbe0e332654-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83659e7b-5ef2-4128-9a9e-4cbe0e332654" (UID: "83659e7b-5ef2-4128-9a9e-4cbe0e332654"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.554561 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-242v6"] Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.607712 4968 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.607762 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtcfw\" (UniqueName: \"kubernetes.io/projected/26f10928-44af-4829-b9f7-1fe382373a48-kube-api-access-jtcfw\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.607773 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83659e7b-5ef2-4128-9a9e-4cbe0e332654-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.607782 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f10928-44af-4829-b9f7-1fe382373a48-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.607790 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w7n5\" (UniqueName: \"kubernetes.io/projected/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-kube-api-access-9w7n5\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.607799 4968 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc03d1b2-f3f3-4b18-b108-e1a166a6a250-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.636736 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26f10928-44af-4829-b9f7-1fe382373a48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26f10928-44af-4829-b9f7-1fe382373a48" (UID: "26f10928-44af-4829-b9f7-1fe382373a48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.710181 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f10928-44af-4829-b9f7-1fe382373a48-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.797559 4968 generic.go:334] "Generic (PLEG): container finished" podID="deeab883-52a5-49df-9b71-e3927cf0f2ea" containerID="53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455" exitCode=0 Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.797613 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4r24p" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.797610 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4r24p" event={"ID":"deeab883-52a5-49df-9b71-e3927cf0f2ea","Type":"ContainerDied","Data":"53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455"} Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.798189 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4r24p" event={"ID":"deeab883-52a5-49df-9b71-e3927cf0f2ea","Type":"ContainerDied","Data":"406ebe4186c7b5a8e378089fa5a0d79826d63c40a1fa5eb7f510908fe02af71d"} Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.798222 4968 scope.go:117] "RemoveContainer" containerID="53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.800986 4968 generic.go:334] "Generic (PLEG): container finished" podID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" containerID="86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e" exitCode=0 Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.801059 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f4xk" event={"ID":"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5","Type":"ContainerDied","Data":"86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e"} Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.801095 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5f4xk" event={"ID":"48c11632-e38d-4cd3-8ac6-9ef42d55b0e5","Type":"ContainerDied","Data":"fb08b8fcc9b0a35b22b7901248aaf447dee4756881117a6e336eeb334523333a"} Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.801099 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5f4xk" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.805095 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-242v6" event={"ID":"e2eb1519-86a1-496a-ace0-ee997d43287c","Type":"ContainerStarted","Data":"8b92f4d4a4d165d68df45bf8f3bdca45f1382bdb443a07f94fb7de1c987384ee"} Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.805339 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.805354 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-242v6" event={"ID":"e2eb1519-86a1-496a-ace0-ee997d43287c","Type":"ContainerStarted","Data":"0407e51528ecdee5ad1d90ab445792d214c378865934706b3f3a96aad5b45ac6"} Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.806529 4968 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-242v6 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" start-of-body= Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.806582 4968 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-242v6" podUID="e2eb1519-86a1-496a-ace0-ee997d43287c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.807837 4968 generic.go:334] "Generic (PLEG): container finished" podID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" containerID="f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a" exitCode=0 Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.807885 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkb2r" event={"ID":"83659e7b-5ef2-4128-9a9e-4cbe0e332654","Type":"ContainerDied","Data":"f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a"} Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.807905 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkb2r" event={"ID":"83659e7b-5ef2-4128-9a9e-4cbe0e332654","Type":"ContainerDied","Data":"275972af3588df7723d415062506c3753ddd7a9f6bf1e912efc48ada3a7926a1"} Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.807969 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkb2r" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.809390 4968 generic.go:334] "Generic (PLEG): container finished" podID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" containerID="d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603" exitCode=0 Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.809437 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" event={"ID":"cc03d1b2-f3f3-4b18-b108-e1a166a6a250","Type":"ContainerDied","Data":"d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603"} Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.809458 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" event={"ID":"cc03d1b2-f3f3-4b18-b108-e1a166a6a250","Type":"ContainerDied","Data":"e58e337ace8777153cb1a3407c99e2c459678535ffd1e4d58842bd61bb53440d"} Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.809513 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dnmb8" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.817249 4968 generic.go:334] "Generic (PLEG): container finished" podID="26f10928-44af-4829-b9f7-1fe382373a48" containerID="8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547" exitCode=0 Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.817288 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrbbc" event={"ID":"26f10928-44af-4829-b9f7-1fe382373a48","Type":"ContainerDied","Data":"8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547"} Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.817311 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrbbc" event={"ID":"26f10928-44af-4829-b9f7-1fe382373a48","Type":"ContainerDied","Data":"1a3d5a4346d54cf5dcb62bdf6eee8f9d1fde9e68130872579c185114fb84b28d"} Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.817366 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mrbbc" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.841247 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-242v6" podStartSLOduration=1.8412307110000001 podStartE2EDuration="1.841230711s" podCreationTimestamp="2026-02-18 15:27:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:27:38.836350404 +0000 UTC m=+358.221795276" watchObservedRunningTime="2026-02-18 15:27:38.841230711 +0000 UTC m=+358.226675573" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.863628 4968 scope.go:117] "RemoveContainer" containerID="27a3f09f40054a3959682f11d2b0d7f37b99e3bb61270998dff67cfadb2a96e4" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.883030 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f4xk"] Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.891223 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5f4xk"] Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.894828 4968 scope.go:117] "RemoveContainer" containerID="522783a09d37975d4c568a594147e1c9b7745f88f09af1bf1ee014ae2ff1cfbd" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.896819 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hkb2r"] Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.903980 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hkb2r"] Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.910988 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dnmb8"] Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.914202 4968 scope.go:117] "RemoveContainer" containerID="53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455" Feb 18 15:27:38 crc kubenswrapper[4968]: E0218 15:27:38.914657 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455\": container with ID starting with 53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455 not found: ID does not exist" containerID="53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.914693 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455"} err="failed to get container status \"53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455\": rpc error: code = NotFound desc = could not find container \"53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455\": container with ID starting with 53274fe6c64b0941b544e288e7b59f16651de77da19f86c3ba725d4a2e85a455 not found: ID does not exist" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.914719 4968 scope.go:117] "RemoveContainer" containerID="27a3f09f40054a3959682f11d2b0d7f37b99e3bb61270998dff67cfadb2a96e4" Feb 18 15:27:38 crc kubenswrapper[4968]: E0218 15:27:38.915067 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27a3f09f40054a3959682f11d2b0d7f37b99e3bb61270998dff67cfadb2a96e4\": container with ID starting with 27a3f09f40054a3959682f11d2b0d7f37b99e3bb61270998dff67cfadb2a96e4 not found: ID does not exist" containerID="27a3f09f40054a3959682f11d2b0d7f37b99e3bb61270998dff67cfadb2a96e4" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.915098 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27a3f09f40054a3959682f11d2b0d7f37b99e3bb61270998dff67cfadb2a96e4"} err="failed to get container status \"27a3f09f40054a3959682f11d2b0d7f37b99e3bb61270998dff67cfadb2a96e4\": rpc error: code = NotFound desc = could not find container \"27a3f09f40054a3959682f11d2b0d7f37b99e3bb61270998dff67cfadb2a96e4\": container with ID starting with 27a3f09f40054a3959682f11d2b0d7f37b99e3bb61270998dff67cfadb2a96e4 not found: ID does not exist" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.915116 4968 scope.go:117] "RemoveContainer" containerID="522783a09d37975d4c568a594147e1c9b7745f88f09af1bf1ee014ae2ff1cfbd" Feb 18 15:27:38 crc kubenswrapper[4968]: E0218 15:27:38.915457 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"522783a09d37975d4c568a594147e1c9b7745f88f09af1bf1ee014ae2ff1cfbd\": container with ID starting with 522783a09d37975d4c568a594147e1c9b7745f88f09af1bf1ee014ae2ff1cfbd not found: ID does not exist" containerID="522783a09d37975d4c568a594147e1c9b7745f88f09af1bf1ee014ae2ff1cfbd" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.915495 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"522783a09d37975d4c568a594147e1c9b7745f88f09af1bf1ee014ae2ff1cfbd"} err="failed to get container status \"522783a09d37975d4c568a594147e1c9b7745f88f09af1bf1ee014ae2ff1cfbd\": rpc error: code = NotFound desc = could not find container \"522783a09d37975d4c568a594147e1c9b7745f88f09af1bf1ee014ae2ff1cfbd\": container with ID starting with 522783a09d37975d4c568a594147e1c9b7745f88f09af1bf1ee014ae2ff1cfbd not found: ID does not exist" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.915514 4968 scope.go:117] "RemoveContainer" containerID="86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.922127 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dnmb8"] Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.931390 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4r24p"] Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.935609 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4r24p"] Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.940350 4968 scope.go:117] "RemoveContainer" containerID="4835c97fbcf6aee48a483c3f83bdde3894eaf7239be9351d9178fb0ff9989865" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.940415 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mrbbc"] Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.945264 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mrbbc"] Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.959069 4968 scope.go:117] "RemoveContainer" containerID="dac10d240dcacfe95fde80436e862a7d7acb55b2f41c492f593261598039ea32" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.979843 4968 scope.go:117] "RemoveContainer" containerID="86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e" Feb 18 15:27:38 crc kubenswrapper[4968]: E0218 15:27:38.980354 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e\": container with ID starting with 86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e not found: ID does not exist" containerID="86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.980398 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e"} err="failed to get container status \"86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e\": rpc error: code = NotFound desc = could not find container \"86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e\": container with ID starting with 86dec471f2d02330cdfd2337004b0fc08224628ff9dbf4d0ec2cea5702e9858e not found: ID does not exist" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.980445 4968 scope.go:117] "RemoveContainer" containerID="4835c97fbcf6aee48a483c3f83bdde3894eaf7239be9351d9178fb0ff9989865" Feb 18 15:27:38 crc kubenswrapper[4968]: E0218 15:27:38.980737 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4835c97fbcf6aee48a483c3f83bdde3894eaf7239be9351d9178fb0ff9989865\": container with ID starting with 4835c97fbcf6aee48a483c3f83bdde3894eaf7239be9351d9178fb0ff9989865 not found: ID does not exist" containerID="4835c97fbcf6aee48a483c3f83bdde3894eaf7239be9351d9178fb0ff9989865" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.980795 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4835c97fbcf6aee48a483c3f83bdde3894eaf7239be9351d9178fb0ff9989865"} err="failed to get container status \"4835c97fbcf6aee48a483c3f83bdde3894eaf7239be9351d9178fb0ff9989865\": rpc error: code = NotFound desc = could not find container \"4835c97fbcf6aee48a483c3f83bdde3894eaf7239be9351d9178fb0ff9989865\": container with ID starting with 4835c97fbcf6aee48a483c3f83bdde3894eaf7239be9351d9178fb0ff9989865 not found: ID does not exist" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.980820 4968 scope.go:117] "RemoveContainer" containerID="dac10d240dcacfe95fde80436e862a7d7acb55b2f41c492f593261598039ea32" Feb 18 15:27:38 crc kubenswrapper[4968]: E0218 15:27:38.981099 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dac10d240dcacfe95fde80436e862a7d7acb55b2f41c492f593261598039ea32\": container with ID starting with dac10d240dcacfe95fde80436e862a7d7acb55b2f41c492f593261598039ea32 not found: ID does not exist" containerID="dac10d240dcacfe95fde80436e862a7d7acb55b2f41c492f593261598039ea32" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.981126 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dac10d240dcacfe95fde80436e862a7d7acb55b2f41c492f593261598039ea32"} err="failed to get container status \"dac10d240dcacfe95fde80436e862a7d7acb55b2f41c492f593261598039ea32\": rpc error: code = NotFound desc = could not find container \"dac10d240dcacfe95fde80436e862a7d7acb55b2f41c492f593261598039ea32\": container with ID starting with dac10d240dcacfe95fde80436e862a7d7acb55b2f41c492f593261598039ea32 not found: ID does not exist" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.981141 4968 scope.go:117] "RemoveContainer" containerID="f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a" Feb 18 15:27:38 crc kubenswrapper[4968]: I0218 15:27:38.997698 4968 scope.go:117] "RemoveContainer" containerID="370174d7c6b8a5d86f2b039b1495819498c0ad720d184695a683709b27914337" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.016448 4968 scope.go:117] "RemoveContainer" containerID="9c3e4b13a9993481130a68da0b12502fc4d2928ba35bcda91db9b6119ec6e067" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.037253 4968 scope.go:117] "RemoveContainer" containerID="f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a" Feb 18 15:27:39 crc kubenswrapper[4968]: E0218 15:27:39.037765 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a\": container with ID starting with f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a not found: ID does not exist" containerID="f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.037816 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a"} err="failed to get container status \"f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a\": rpc error: code = NotFound desc = could not find container \"f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a\": container with ID starting with f2c56d5924ea3c4173cab2823eab37798c079b2b383dc6c6cda0286c674e161a not found: ID does not exist" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.037852 4968 scope.go:117] "RemoveContainer" containerID="370174d7c6b8a5d86f2b039b1495819498c0ad720d184695a683709b27914337" Feb 18 15:27:39 crc kubenswrapper[4968]: E0218 15:27:39.038180 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"370174d7c6b8a5d86f2b039b1495819498c0ad720d184695a683709b27914337\": container with ID starting with 370174d7c6b8a5d86f2b039b1495819498c0ad720d184695a683709b27914337 not found: ID does not exist" containerID="370174d7c6b8a5d86f2b039b1495819498c0ad720d184695a683709b27914337" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.038215 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"370174d7c6b8a5d86f2b039b1495819498c0ad720d184695a683709b27914337"} err="failed to get container status \"370174d7c6b8a5d86f2b039b1495819498c0ad720d184695a683709b27914337\": rpc error: code = NotFound desc = could not find container \"370174d7c6b8a5d86f2b039b1495819498c0ad720d184695a683709b27914337\": container with ID starting with 370174d7c6b8a5d86f2b039b1495819498c0ad720d184695a683709b27914337 not found: ID does not exist" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.038238 4968 scope.go:117] "RemoveContainer" containerID="9c3e4b13a9993481130a68da0b12502fc4d2928ba35bcda91db9b6119ec6e067" Feb 18 15:27:39 crc kubenswrapper[4968]: E0218 15:27:39.038625 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c3e4b13a9993481130a68da0b12502fc4d2928ba35bcda91db9b6119ec6e067\": container with ID starting with 9c3e4b13a9993481130a68da0b12502fc4d2928ba35bcda91db9b6119ec6e067 not found: ID does not exist" containerID="9c3e4b13a9993481130a68da0b12502fc4d2928ba35bcda91db9b6119ec6e067" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.038702 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c3e4b13a9993481130a68da0b12502fc4d2928ba35bcda91db9b6119ec6e067"} err="failed to get container status \"9c3e4b13a9993481130a68da0b12502fc4d2928ba35bcda91db9b6119ec6e067\": rpc error: code = NotFound desc = could not find container \"9c3e4b13a9993481130a68da0b12502fc4d2928ba35bcda91db9b6119ec6e067\": container with ID starting with 9c3e4b13a9993481130a68da0b12502fc4d2928ba35bcda91db9b6119ec6e067 not found: ID does not exist" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.038785 4968 scope.go:117] "RemoveContainer" containerID="d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.067322 4968 scope.go:117] "RemoveContainer" containerID="a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.085190 4968 scope.go:117] "RemoveContainer" containerID="d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603" Feb 18 15:27:39 crc kubenswrapper[4968]: E0218 15:27:39.086010 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603\": container with ID starting with d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603 not found: ID does not exist" containerID="d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.086082 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603"} err="failed to get container status \"d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603\": rpc error: code = NotFound desc = could not find container \"d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603\": container with ID starting with d13ed57665ab80816fcfd125ea577bf54ea7b2658ee770e47c0d66dae0f92603 not found: ID does not exist" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.086123 4968 scope.go:117] "RemoveContainer" containerID="a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212" Feb 18 15:27:39 crc kubenswrapper[4968]: E0218 15:27:39.087069 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212\": container with ID starting with a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212 not found: ID does not exist" containerID="a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.087113 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212"} err="failed to get container status \"a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212\": rpc error: code = NotFound desc = could not find container \"a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212\": container with ID starting with a534d2685ceaca43587636b8c6b1999380efebd814dd73357cb42f91599ae212 not found: ID does not exist" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.087144 4968 scope.go:117] "RemoveContainer" containerID="8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.109238 4968 scope.go:117] "RemoveContainer" containerID="09a54638221a06794178f0003151e08b499b0459a39c3b6afb46f09037f4f324" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.128679 4968 scope.go:117] "RemoveContainer" containerID="0784ec62ed1129aaa9288f7d38add433d93111e93fc19a588beacf63718b9a3c" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.143921 4968 scope.go:117] "RemoveContainer" containerID="8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547" Feb 18 15:27:39 crc kubenswrapper[4968]: E0218 15:27:39.144700 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547\": container with ID starting with 8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547 not found: ID does not exist" containerID="8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.144763 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547"} err="failed to get container status \"8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547\": rpc error: code = NotFound desc = could not find container \"8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547\": container with ID starting with 8a61cbb493ddd9ea42090c31c619e65e630a635a2071dd94af1878ef198c0547 not found: ID does not exist" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.144792 4968 scope.go:117] "RemoveContainer" containerID="09a54638221a06794178f0003151e08b499b0459a39c3b6afb46f09037f4f324" Feb 18 15:27:39 crc kubenswrapper[4968]: E0218 15:27:39.145179 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09a54638221a06794178f0003151e08b499b0459a39c3b6afb46f09037f4f324\": container with ID starting with 09a54638221a06794178f0003151e08b499b0459a39c3b6afb46f09037f4f324 not found: ID does not exist" containerID="09a54638221a06794178f0003151e08b499b0459a39c3b6afb46f09037f4f324" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.145241 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09a54638221a06794178f0003151e08b499b0459a39c3b6afb46f09037f4f324"} err="failed to get container status \"09a54638221a06794178f0003151e08b499b0459a39c3b6afb46f09037f4f324\": rpc error: code = NotFound desc = could not find container \"09a54638221a06794178f0003151e08b499b0459a39c3b6afb46f09037f4f324\": container with ID starting with 09a54638221a06794178f0003151e08b499b0459a39c3b6afb46f09037f4f324 not found: ID does not exist" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.145269 4968 scope.go:117] "RemoveContainer" containerID="0784ec62ed1129aaa9288f7d38add433d93111e93fc19a588beacf63718b9a3c" Feb 18 15:27:39 crc kubenswrapper[4968]: E0218 15:27:39.145623 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0784ec62ed1129aaa9288f7d38add433d93111e93fc19a588beacf63718b9a3c\": container with ID starting with 0784ec62ed1129aaa9288f7d38add433d93111e93fc19a588beacf63718b9a3c not found: ID does not exist" containerID="0784ec62ed1129aaa9288f7d38add433d93111e93fc19a588beacf63718b9a3c" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.145655 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0784ec62ed1129aaa9288f7d38add433d93111e93fc19a588beacf63718b9a3c"} err="failed to get container status \"0784ec62ed1129aaa9288f7d38add433d93111e93fc19a588beacf63718b9a3c\": rpc error: code = NotFound desc = could not find container \"0784ec62ed1129aaa9288f7d38add433d93111e93fc19a588beacf63718b9a3c\": container with ID starting with 0784ec62ed1129aaa9288f7d38add433d93111e93fc19a588beacf63718b9a3c not found: ID does not exist" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.242880 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26f10928-44af-4829-b9f7-1fe382373a48" path="/var/lib/kubelet/pods/26f10928-44af-4829-b9f7-1fe382373a48/volumes" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.244135 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" path="/var/lib/kubelet/pods/48c11632-e38d-4cd3-8ac6-9ef42d55b0e5/volumes" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.245288 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" path="/var/lib/kubelet/pods/83659e7b-5ef2-4128-9a9e-4cbe0e332654/volumes" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.247433 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" path="/var/lib/kubelet/pods/cc03d1b2-f3f3-4b18-b108-e1a166a6a250/volumes" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.248373 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deeab883-52a5-49df-9b71-e3927cf0f2ea" path="/var/lib/kubelet/pods/deeab883-52a5-49df-9b71-e3927cf0f2ea/volumes" Feb 18 15:27:39 crc kubenswrapper[4968]: I0218 15:27:39.834420 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-242v6" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.107998 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8m6ss"] Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108676 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" containerName="extract-utilities" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108694 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" containerName="extract-utilities" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108709 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f10928-44af-4829-b9f7-1fe382373a48" containerName="extract-content" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108716 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f10928-44af-4829-b9f7-1fe382373a48" containerName="extract-content" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108731 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" containerName="registry-server" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108737 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" containerName="registry-server" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108745 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeab883-52a5-49df-9b71-e3927cf0f2ea" containerName="extract-utilities" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108768 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeab883-52a5-49df-9b71-e3927cf0f2ea" containerName="extract-utilities" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108778 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" containerName="marketplace-operator" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108783 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" containerName="marketplace-operator" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108790 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f10928-44af-4829-b9f7-1fe382373a48" containerName="registry-server" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108796 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f10928-44af-4829-b9f7-1fe382373a48" containerName="registry-server" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108806 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeab883-52a5-49df-9b71-e3927cf0f2ea" containerName="extract-content" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108815 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeab883-52a5-49df-9b71-e3927cf0f2ea" containerName="extract-content" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108823 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" containerName="extract-content" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108830 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" containerName="extract-content" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108841 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeab883-52a5-49df-9b71-e3927cf0f2ea" containerName="registry-server" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108846 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeab883-52a5-49df-9b71-e3927cf0f2ea" containerName="registry-server" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108856 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" containerName="extract-content" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108864 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" containerName="extract-content" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108871 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" containerName="registry-server" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108876 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" containerName="registry-server" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108885 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f10928-44af-4829-b9f7-1fe382373a48" containerName="extract-utilities" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108891 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f10928-44af-4829-b9f7-1fe382373a48" containerName="extract-utilities" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108899 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" containerName="marketplace-operator" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108905 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" containerName="marketplace-operator" Feb 18 15:27:40 crc kubenswrapper[4968]: E0218 15:27:40.108914 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" containerName="extract-utilities" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.108920 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" containerName="extract-utilities" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.109028 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" containerName="marketplace-operator" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.109040 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f10928-44af-4829-b9f7-1fe382373a48" containerName="registry-server" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.109055 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="83659e7b-5ef2-4128-9a9e-4cbe0e332654" containerName="registry-server" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.109064 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeab883-52a5-49df-9b71-e3927cf0f2ea" containerName="registry-server" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.109073 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c11632-e38d-4cd3-8ac6-9ef42d55b0e5" containerName="registry-server" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.109295 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc03d1b2-f3f3-4b18-b108-e1a166a6a250" containerName="marketplace-operator" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.109999 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.112461 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.122106 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8m6ss"] Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.231007 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpp2b\" (UniqueName: \"kubernetes.io/projected/c09f5ccb-47bd-46ca-8f51-1fc196f396ee-kube-api-access-hpp2b\") pod \"redhat-marketplace-8m6ss\" (UID: \"c09f5ccb-47bd-46ca-8f51-1fc196f396ee\") " pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.231379 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c09f5ccb-47bd-46ca-8f51-1fc196f396ee-catalog-content\") pod \"redhat-marketplace-8m6ss\" (UID: \"c09f5ccb-47bd-46ca-8f51-1fc196f396ee\") " pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.231524 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c09f5ccb-47bd-46ca-8f51-1fc196f396ee-utilities\") pod \"redhat-marketplace-8m6ss\" (UID: \"c09f5ccb-47bd-46ca-8f51-1fc196f396ee\") " pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.301814 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9v9wz"] Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.302941 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.307596 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.324922 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9v9wz"] Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.332509 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c09f5ccb-47bd-46ca-8f51-1fc196f396ee-catalog-content\") pod \"redhat-marketplace-8m6ss\" (UID: \"c09f5ccb-47bd-46ca-8f51-1fc196f396ee\") " pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.332741 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c09f5ccb-47bd-46ca-8f51-1fc196f396ee-utilities\") pod \"redhat-marketplace-8m6ss\" (UID: \"c09f5ccb-47bd-46ca-8f51-1fc196f396ee\") " pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.332890 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpp2b\" (UniqueName: \"kubernetes.io/projected/c09f5ccb-47bd-46ca-8f51-1fc196f396ee-kube-api-access-hpp2b\") pod \"redhat-marketplace-8m6ss\" (UID: \"c09f5ccb-47bd-46ca-8f51-1fc196f396ee\") " pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.333322 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c09f5ccb-47bd-46ca-8f51-1fc196f396ee-catalog-content\") pod \"redhat-marketplace-8m6ss\" (UID: \"c09f5ccb-47bd-46ca-8f51-1fc196f396ee\") " pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.333794 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c09f5ccb-47bd-46ca-8f51-1fc196f396ee-utilities\") pod \"redhat-marketplace-8m6ss\" (UID: \"c09f5ccb-47bd-46ca-8f51-1fc196f396ee\") " pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.366518 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpp2b\" (UniqueName: \"kubernetes.io/projected/c09f5ccb-47bd-46ca-8f51-1fc196f396ee-kube-api-access-hpp2b\") pod \"redhat-marketplace-8m6ss\" (UID: \"c09f5ccb-47bd-46ca-8f51-1fc196f396ee\") " pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.434511 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3c68c1d-979b-4059-96bb-8c1415fb9776-utilities\") pod \"redhat-operators-9v9wz\" (UID: \"d3c68c1d-979b-4059-96bb-8c1415fb9776\") " pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.434579 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xth4w\" (UniqueName: \"kubernetes.io/projected/d3c68c1d-979b-4059-96bb-8c1415fb9776-kube-api-access-xth4w\") pod \"redhat-operators-9v9wz\" (UID: \"d3c68c1d-979b-4059-96bb-8c1415fb9776\") " pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.434646 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3c68c1d-979b-4059-96bb-8c1415fb9776-catalog-content\") pod \"redhat-operators-9v9wz\" (UID: \"d3c68c1d-979b-4059-96bb-8c1415fb9776\") " pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.440360 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.535923 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3c68c1d-979b-4059-96bb-8c1415fb9776-utilities\") pod \"redhat-operators-9v9wz\" (UID: \"d3c68c1d-979b-4059-96bb-8c1415fb9776\") " pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.536026 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xth4w\" (UniqueName: \"kubernetes.io/projected/d3c68c1d-979b-4059-96bb-8c1415fb9776-kube-api-access-xth4w\") pod \"redhat-operators-9v9wz\" (UID: \"d3c68c1d-979b-4059-96bb-8c1415fb9776\") " pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.536985 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3c68c1d-979b-4059-96bb-8c1415fb9776-catalog-content\") pod \"redhat-operators-9v9wz\" (UID: \"d3c68c1d-979b-4059-96bb-8c1415fb9776\") " pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.537244 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3c68c1d-979b-4059-96bb-8c1415fb9776-utilities\") pod \"redhat-operators-9v9wz\" (UID: \"d3c68c1d-979b-4059-96bb-8c1415fb9776\") " pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.537481 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3c68c1d-979b-4059-96bb-8c1415fb9776-catalog-content\") pod \"redhat-operators-9v9wz\" (UID: \"d3c68c1d-979b-4059-96bb-8c1415fb9776\") " pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.552406 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xth4w\" (UniqueName: \"kubernetes.io/projected/d3c68c1d-979b-4059-96bb-8c1415fb9776-kube-api-access-xth4w\") pod \"redhat-operators-9v9wz\" (UID: \"d3c68c1d-979b-4059-96bb-8c1415fb9776\") " pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.624316 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.819784 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9v9wz"] Feb 18 15:27:40 crc kubenswrapper[4968]: W0218 15:27:40.833306 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3c68c1d_979b_4059_96bb_8c1415fb9776.slice/crio-f9e21aa7dbf46e9e6047a5a8840d655228831acf92393bc58b2179698d93b214 WatchSource:0}: Error finding container f9e21aa7dbf46e9e6047a5a8840d655228831acf92393bc58b2179698d93b214: Status 404 returned error can't find the container with id f9e21aa7dbf46e9e6047a5a8840d655228831acf92393bc58b2179698d93b214 Feb 18 15:27:40 crc kubenswrapper[4968]: I0218 15:27:40.864406 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8m6ss"] Feb 18 15:27:40 crc kubenswrapper[4968]: W0218 15:27:40.872852 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc09f5ccb_47bd_46ca_8f51_1fc196f396ee.slice/crio-59428065bc85e67f7d6b37353e5c00869cd543b82aeaca864cee17da3691bcce WatchSource:0}: Error finding container 59428065bc85e67f7d6b37353e5c00869cd543b82aeaca864cee17da3691bcce: Status 404 returned error can't find the container with id 59428065bc85e67f7d6b37353e5c00869cd543b82aeaca864cee17da3691bcce Feb 18 15:27:41 crc kubenswrapper[4968]: I0218 15:27:41.846063 4968 generic.go:334] "Generic (PLEG): container finished" podID="d3c68c1d-979b-4059-96bb-8c1415fb9776" containerID="eeb56d4fee21f21ea4d0bd26231a6abff321639e86438c99b46ab948265b489b" exitCode=0 Feb 18 15:27:41 crc kubenswrapper[4968]: I0218 15:27:41.846192 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9v9wz" event={"ID":"d3c68c1d-979b-4059-96bb-8c1415fb9776","Type":"ContainerDied","Data":"eeb56d4fee21f21ea4d0bd26231a6abff321639e86438c99b46ab948265b489b"} Feb 18 15:27:41 crc kubenswrapper[4968]: I0218 15:27:41.846351 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9v9wz" event={"ID":"d3c68c1d-979b-4059-96bb-8c1415fb9776","Type":"ContainerStarted","Data":"f9e21aa7dbf46e9e6047a5a8840d655228831acf92393bc58b2179698d93b214"} Feb 18 15:27:41 crc kubenswrapper[4968]: I0218 15:27:41.849237 4968 generic.go:334] "Generic (PLEG): container finished" podID="c09f5ccb-47bd-46ca-8f51-1fc196f396ee" containerID="4a24ef72e9f938bd1cd06fdfe332fb39de4bbf8def4b316ddd1f8fd981c3528f" exitCode=0 Feb 18 15:27:41 crc kubenswrapper[4968]: I0218 15:27:41.849277 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8m6ss" event={"ID":"c09f5ccb-47bd-46ca-8f51-1fc196f396ee","Type":"ContainerDied","Data":"4a24ef72e9f938bd1cd06fdfe332fb39de4bbf8def4b316ddd1f8fd981c3528f"} Feb 18 15:27:41 crc kubenswrapper[4968]: I0218 15:27:41.849303 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8m6ss" event={"ID":"c09f5ccb-47bd-46ca-8f51-1fc196f396ee","Type":"ContainerStarted","Data":"59428065bc85e67f7d6b37353e5c00869cd543b82aeaca864cee17da3691bcce"} Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.497850 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-blvt2"] Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.499194 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.503764 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.510549 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-blvt2"] Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.675018 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3589f4c2-2e33-4458-ae09-8844e66b94fd-catalog-content\") pod \"certified-operators-blvt2\" (UID: \"3589f4c2-2e33-4458-ae09-8844e66b94fd\") " pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.675080 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kbqp\" (UniqueName: \"kubernetes.io/projected/3589f4c2-2e33-4458-ae09-8844e66b94fd-kube-api-access-8kbqp\") pod \"certified-operators-blvt2\" (UID: \"3589f4c2-2e33-4458-ae09-8844e66b94fd\") " pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.675106 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3589f4c2-2e33-4458-ae09-8844e66b94fd-utilities\") pod \"certified-operators-blvt2\" (UID: \"3589f4c2-2e33-4458-ae09-8844e66b94fd\") " pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.695606 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4kpft"] Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.697046 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.700802 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.711265 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4kpft"] Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.776107 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kbqp\" (UniqueName: \"kubernetes.io/projected/3589f4c2-2e33-4458-ae09-8844e66b94fd-kube-api-access-8kbqp\") pod \"certified-operators-blvt2\" (UID: \"3589f4c2-2e33-4458-ae09-8844e66b94fd\") " pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.776159 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3589f4c2-2e33-4458-ae09-8844e66b94fd-utilities\") pod \"certified-operators-blvt2\" (UID: \"3589f4c2-2e33-4458-ae09-8844e66b94fd\") " pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.776212 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3589f4c2-2e33-4458-ae09-8844e66b94fd-catalog-content\") pod \"certified-operators-blvt2\" (UID: \"3589f4c2-2e33-4458-ae09-8844e66b94fd\") " pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.776615 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3589f4c2-2e33-4458-ae09-8844e66b94fd-catalog-content\") pod \"certified-operators-blvt2\" (UID: \"3589f4c2-2e33-4458-ae09-8844e66b94fd\") " pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.776832 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3589f4c2-2e33-4458-ae09-8844e66b94fd-utilities\") pod \"certified-operators-blvt2\" (UID: \"3589f4c2-2e33-4458-ae09-8844e66b94fd\") " pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.794119 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kbqp\" (UniqueName: \"kubernetes.io/projected/3589f4c2-2e33-4458-ae09-8844e66b94fd-kube-api-access-8kbqp\") pod \"certified-operators-blvt2\" (UID: \"3589f4c2-2e33-4458-ae09-8844e66b94fd\") " pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.822819 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.869441 4968 generic.go:334] "Generic (PLEG): container finished" podID="c09f5ccb-47bd-46ca-8f51-1fc196f396ee" containerID="86d3761b942d547694d18defded3f711ef1000bad0b895aa50f448c7fb1adf4d" exitCode=0 Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.869557 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8m6ss" event={"ID":"c09f5ccb-47bd-46ca-8f51-1fc196f396ee","Type":"ContainerDied","Data":"86d3761b942d547694d18defded3f711ef1000bad0b895aa50f448c7fb1adf4d"} Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.878025 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8w6f\" (UniqueName: \"kubernetes.io/projected/5391b9a3-b12d-4c69-9d11-b29cbc44b531-kube-api-access-g8w6f\") pod \"community-operators-4kpft\" (UID: \"5391b9a3-b12d-4c69-9d11-b29cbc44b531\") " pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.878098 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5391b9a3-b12d-4c69-9d11-b29cbc44b531-utilities\") pod \"community-operators-4kpft\" (UID: \"5391b9a3-b12d-4c69-9d11-b29cbc44b531\") " pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.878172 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5391b9a3-b12d-4c69-9d11-b29cbc44b531-catalog-content\") pod \"community-operators-4kpft\" (UID: \"5391b9a3-b12d-4c69-9d11-b29cbc44b531\") " pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.882330 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9v9wz" event={"ID":"d3c68c1d-979b-4059-96bb-8c1415fb9776","Type":"ContainerStarted","Data":"179574daa388c576258add2f3fc3c6757325e019284e9f1e232e82813a558272"} Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.979696 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5391b9a3-b12d-4c69-9d11-b29cbc44b531-utilities\") pod \"community-operators-4kpft\" (UID: \"5391b9a3-b12d-4c69-9d11-b29cbc44b531\") " pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.980112 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5391b9a3-b12d-4c69-9d11-b29cbc44b531-catalog-content\") pod \"community-operators-4kpft\" (UID: \"5391b9a3-b12d-4c69-9d11-b29cbc44b531\") " pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.980171 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8w6f\" (UniqueName: \"kubernetes.io/projected/5391b9a3-b12d-4c69-9d11-b29cbc44b531-kube-api-access-g8w6f\") pod \"community-operators-4kpft\" (UID: \"5391b9a3-b12d-4c69-9d11-b29cbc44b531\") " pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.980619 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5391b9a3-b12d-4c69-9d11-b29cbc44b531-utilities\") pod \"community-operators-4kpft\" (UID: \"5391b9a3-b12d-4c69-9d11-b29cbc44b531\") " pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.981204 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5391b9a3-b12d-4c69-9d11-b29cbc44b531-catalog-content\") pod \"community-operators-4kpft\" (UID: \"5391b9a3-b12d-4c69-9d11-b29cbc44b531\") " pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:42 crc kubenswrapper[4968]: I0218 15:27:42.998996 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8w6f\" (UniqueName: \"kubernetes.io/projected/5391b9a3-b12d-4c69-9d11-b29cbc44b531-kube-api-access-g8w6f\") pod \"community-operators-4kpft\" (UID: \"5391b9a3-b12d-4c69-9d11-b29cbc44b531\") " pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.024620 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.238407 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-blvt2"] Feb 18 15:27:43 crc kubenswrapper[4968]: W0218 15:27:43.239119 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3589f4c2_2e33_4458_ae09_8844e66b94fd.slice/crio-7d1b57b0209cc4702f6f105e35e3bace4cceb09474babaf0c17d9da051dec29b WatchSource:0}: Error finding container 7d1b57b0209cc4702f6f105e35e3bace4cceb09474babaf0c17d9da051dec29b: Status 404 returned error can't find the container with id 7d1b57b0209cc4702f6f105e35e3bace4cceb09474babaf0c17d9da051dec29b Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.434555 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4kpft"] Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.897409 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8m6ss" event={"ID":"c09f5ccb-47bd-46ca-8f51-1fc196f396ee","Type":"ContainerStarted","Data":"1820acd97a7b01f147545895593455b1431d2f5817b12a3e56024ba6f1ef997f"} Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.899352 4968 generic.go:334] "Generic (PLEG): container finished" podID="d3c68c1d-979b-4059-96bb-8c1415fb9776" containerID="179574daa388c576258add2f3fc3c6757325e019284e9f1e232e82813a558272" exitCode=0 Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.899390 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9v9wz" event={"ID":"d3c68c1d-979b-4059-96bb-8c1415fb9776","Type":"ContainerDied","Data":"179574daa388c576258add2f3fc3c6757325e019284e9f1e232e82813a558272"} Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.901731 4968 generic.go:334] "Generic (PLEG): container finished" podID="3589f4c2-2e33-4458-ae09-8844e66b94fd" containerID="71251b4302c5cf239590e9622935b3f481fe6fd508fd1259f8c6614b5645429f" exitCode=0 Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.901781 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blvt2" event={"ID":"3589f4c2-2e33-4458-ae09-8844e66b94fd","Type":"ContainerDied","Data":"71251b4302c5cf239590e9622935b3f481fe6fd508fd1259f8c6614b5645429f"} Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.901818 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blvt2" event={"ID":"3589f4c2-2e33-4458-ae09-8844e66b94fd","Type":"ContainerStarted","Data":"7d1b57b0209cc4702f6f105e35e3bace4cceb09474babaf0c17d9da051dec29b"} Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.904557 4968 generic.go:334] "Generic (PLEG): container finished" podID="5391b9a3-b12d-4c69-9d11-b29cbc44b531" containerID="da0ed0bc2b2670ded89ddc1b1d1caf5cf1d8d862db4d817b99a40bae79b3a83a" exitCode=0 Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.904607 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4kpft" event={"ID":"5391b9a3-b12d-4c69-9d11-b29cbc44b531","Type":"ContainerDied","Data":"da0ed0bc2b2670ded89ddc1b1d1caf5cf1d8d862db4d817b99a40bae79b3a83a"} Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.904639 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4kpft" event={"ID":"5391b9a3-b12d-4c69-9d11-b29cbc44b531","Type":"ContainerStarted","Data":"fefe14910b4e9f6a2befe0d21ede5d36de6b63f2701285abbf966876536c1386"} Feb 18 15:27:43 crc kubenswrapper[4968]: I0218 15:27:43.920372 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8m6ss" podStartSLOduration=2.468863114 podStartE2EDuration="3.9203543s" podCreationTimestamp="2026-02-18 15:27:40 +0000 UTC" firstStartedPulling="2026-02-18 15:27:41.85088674 +0000 UTC m=+361.236331612" lastFinishedPulling="2026-02-18 15:27:43.302377936 +0000 UTC m=+362.687822798" observedRunningTime="2026-02-18 15:27:43.916907844 +0000 UTC m=+363.302352716" watchObservedRunningTime="2026-02-18 15:27:43.9203543 +0000 UTC m=+363.305799162" Feb 18 15:27:44 crc kubenswrapper[4968]: I0218 15:27:44.367853 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:27:44 crc kubenswrapper[4968]: I0218 15:27:44.367911 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:27:44 crc kubenswrapper[4968]: I0218 15:27:44.917901 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blvt2" event={"ID":"3589f4c2-2e33-4458-ae09-8844e66b94fd","Type":"ContainerStarted","Data":"406abc846d20e1c87df6c9153dc08836a60ab3c01e6d33ce02aa6ec08939e0f8"} Feb 18 15:27:44 crc kubenswrapper[4968]: I0218 15:27:44.920896 4968 generic.go:334] "Generic (PLEG): container finished" podID="5391b9a3-b12d-4c69-9d11-b29cbc44b531" containerID="e62b53d689ce43cb0f59bd53a79ededd3f4fbb70248892021e5bb470d903b6c1" exitCode=0 Feb 18 15:27:44 crc kubenswrapper[4968]: I0218 15:27:44.920952 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4kpft" event={"ID":"5391b9a3-b12d-4c69-9d11-b29cbc44b531","Type":"ContainerDied","Data":"e62b53d689ce43cb0f59bd53a79ededd3f4fbb70248892021e5bb470d903b6c1"} Feb 18 15:27:44 crc kubenswrapper[4968]: I0218 15:27:44.930408 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9v9wz" event={"ID":"d3c68c1d-979b-4059-96bb-8c1415fb9776","Type":"ContainerStarted","Data":"0137e757cab2198962840a05a3de784bd956983ad77c50f73c21a64f3dce3a21"} Feb 18 15:27:44 crc kubenswrapper[4968]: I0218 15:27:44.983139 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9v9wz" podStartSLOduration=2.435985586 podStartE2EDuration="4.983120382s" podCreationTimestamp="2026-02-18 15:27:40 +0000 UTC" firstStartedPulling="2026-02-18 15:27:41.848286848 +0000 UTC m=+361.233731710" lastFinishedPulling="2026-02-18 15:27:44.395421644 +0000 UTC m=+363.780866506" observedRunningTime="2026-02-18 15:27:44.978892774 +0000 UTC m=+364.364337636" watchObservedRunningTime="2026-02-18 15:27:44.983120382 +0000 UTC m=+364.368565254" Feb 18 15:27:45 crc kubenswrapper[4968]: E0218 15:27:45.037033 4968 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3589f4c2_2e33_4458_ae09_8844e66b94fd.slice/crio-406abc846d20e1c87df6c9153dc08836a60ab3c01e6d33ce02aa6ec08939e0f8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3589f4c2_2e33_4458_ae09_8844e66b94fd.slice/crio-conmon-406abc846d20e1c87df6c9153dc08836a60ab3c01e6d33ce02aa6ec08939e0f8.scope\": RecentStats: unable to find data in memory cache]" Feb 18 15:27:45 crc kubenswrapper[4968]: I0218 15:27:45.935499 4968 generic.go:334] "Generic (PLEG): container finished" podID="3589f4c2-2e33-4458-ae09-8844e66b94fd" containerID="406abc846d20e1c87df6c9153dc08836a60ab3c01e6d33ce02aa6ec08939e0f8" exitCode=0 Feb 18 15:27:45 crc kubenswrapper[4968]: I0218 15:27:45.935550 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blvt2" event={"ID":"3589f4c2-2e33-4458-ae09-8844e66b94fd","Type":"ContainerDied","Data":"406abc846d20e1c87df6c9153dc08836a60ab3c01e6d33ce02aa6ec08939e0f8"} Feb 18 15:27:45 crc kubenswrapper[4968]: I0218 15:27:45.939940 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4kpft" event={"ID":"5391b9a3-b12d-4c69-9d11-b29cbc44b531","Type":"ContainerStarted","Data":"88e7fc514c1fbd5a69a3d8a4b11f6d39919e861703374719e5433122b76fd22f"} Feb 18 15:27:45 crc kubenswrapper[4968]: I0218 15:27:45.968913 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4kpft" podStartSLOduration=2.5462639339999997 podStartE2EDuration="3.968893974s" podCreationTimestamp="2026-02-18 15:27:42 +0000 UTC" firstStartedPulling="2026-02-18 15:27:43.909171468 +0000 UTC m=+363.294616330" lastFinishedPulling="2026-02-18 15:27:45.331801498 +0000 UTC m=+364.717246370" observedRunningTime="2026-02-18 15:27:45.967627109 +0000 UTC m=+365.353071971" watchObservedRunningTime="2026-02-18 15:27:45.968893974 +0000 UTC m=+365.354338836" Feb 18 15:27:46 crc kubenswrapper[4968]: I0218 15:27:46.946557 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blvt2" event={"ID":"3589f4c2-2e33-4458-ae09-8844e66b94fd","Type":"ContainerStarted","Data":"876ee1c0d2bed1601b3718cf2773363cb64f7d86a43e57ee5ca8532b7dd4af7a"} Feb 18 15:27:46 crc kubenswrapper[4968]: I0218 15:27:46.965464 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-blvt2" podStartSLOduration=2.402143582 podStartE2EDuration="4.965441088s" podCreationTimestamp="2026-02-18 15:27:42 +0000 UTC" firstStartedPulling="2026-02-18 15:27:43.907918783 +0000 UTC m=+363.293363645" lastFinishedPulling="2026-02-18 15:27:46.471216289 +0000 UTC m=+365.856661151" observedRunningTime="2026-02-18 15:27:46.963168214 +0000 UTC m=+366.348613086" watchObservedRunningTime="2026-02-18 15:27:46.965441088 +0000 UTC m=+366.350885950" Feb 18 15:27:50 crc kubenswrapper[4968]: I0218 15:27:50.441292 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:50 crc kubenswrapper[4968]: I0218 15:27:50.442134 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:50 crc kubenswrapper[4968]: I0218 15:27:50.501451 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:50 crc kubenswrapper[4968]: I0218 15:27:50.625346 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:50 crc kubenswrapper[4968]: I0218 15:27:50.625387 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:50 crc kubenswrapper[4968]: I0218 15:27:50.663110 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:51 crc kubenswrapper[4968]: I0218 15:27:51.022072 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8m6ss" Feb 18 15:27:51 crc kubenswrapper[4968]: I0218 15:27:51.025001 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:27:52 crc kubenswrapper[4968]: I0218 15:27:52.823340 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:52 crc kubenswrapper[4968]: I0218 15:27:52.823644 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:52 crc kubenswrapper[4968]: I0218 15:27:52.861655 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:53 crc kubenswrapper[4968]: I0218 15:27:53.024808 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:53 crc kubenswrapper[4968]: I0218 15:27:53.024882 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:53 crc kubenswrapper[4968]: I0218 15:27:53.031112 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-blvt2" Feb 18 15:27:53 crc kubenswrapper[4968]: I0218 15:27:53.081484 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:54 crc kubenswrapper[4968]: I0218 15:27:54.030563 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4kpft" Feb 18 15:27:59 crc kubenswrapper[4968]: I0218 15:27:59.930281 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" podUID="19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" containerName="registry" containerID="cri-o://89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078" gracePeriod=30 Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.284645 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.429695 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.429786 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-registry-certificates\") pod \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.429827 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-bound-sa-token\") pod \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.429889 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-installation-pull-secrets\") pod \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.429906 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwqbc\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-kube-api-access-kwqbc\") pod \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.429942 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-registry-tls\") pod \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.429967 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-trusted-ca\") pod \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.429990 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-ca-trust-extracted\") pod \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\" (UID: \"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2\") " Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.430370 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.430654 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.435873 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.437226 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.438988 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.447161 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.452599 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-kube-api-access-kwqbc" (OuterVolumeSpecName: "kube-api-access-kwqbc") pod "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2"). InnerVolumeSpecName "kube-api-access-kwqbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.458967 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" (UID: "19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.531250 4968 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.531287 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwqbc\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-kube-api-access-kwqbc\") on node \"crc\" DevicePath \"\"" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.531297 4968 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.531306 4968 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.531316 4968 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.531333 4968 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 18 15:28:00 crc kubenswrapper[4968]: I0218 15:28:00.531342 4968 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 18 15:28:01 crc kubenswrapper[4968]: I0218 15:28:01.022697 4968 generic.go:334] "Generic (PLEG): container finished" podID="19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" containerID="89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078" exitCode=0 Feb 18 15:28:01 crc kubenswrapper[4968]: I0218 15:28:01.022738 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" event={"ID":"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2","Type":"ContainerDied","Data":"89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078"} Feb 18 15:28:01 crc kubenswrapper[4968]: I0218 15:28:01.022792 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" event={"ID":"19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2","Type":"ContainerDied","Data":"0fee35f6bd2711ee2dfffb4ab1e34b5a89b04cebc5b4b34bf8aca71759cc639f"} Feb 18 15:28:01 crc kubenswrapper[4968]: I0218 15:28:01.022815 4968 scope.go:117] "RemoveContainer" containerID="89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078" Feb 18 15:28:01 crc kubenswrapper[4968]: I0218 15:28:01.022795 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lcx97" Feb 18 15:28:01 crc kubenswrapper[4968]: I0218 15:28:01.053471 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lcx97"] Feb 18 15:28:01 crc kubenswrapper[4968]: I0218 15:28:01.055092 4968 scope.go:117] "RemoveContainer" containerID="89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078" Feb 18 15:28:01 crc kubenswrapper[4968]: E0218 15:28:01.056433 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078\": container with ID starting with 89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078 not found: ID does not exist" containerID="89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078" Feb 18 15:28:01 crc kubenswrapper[4968]: I0218 15:28:01.056470 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078"} err="failed to get container status \"89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078\": rpc error: code = NotFound desc = could not find container \"89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078\": container with ID starting with 89d9de9808b5d7969e442c4178323c4e1f2d47fa47c0bdc5102138186ded5078 not found: ID does not exist" Feb 18 15:28:01 crc kubenswrapper[4968]: I0218 15:28:01.057802 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lcx97"] Feb 18 15:28:01 crc kubenswrapper[4968]: I0218 15:28:01.238419 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" path="/var/lib/kubelet/pods/19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2/volumes" Feb 18 15:28:14 crc kubenswrapper[4968]: I0218 15:28:14.367119 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:28:14 crc kubenswrapper[4968]: I0218 15:28:14.367614 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:28:44 crc kubenswrapper[4968]: I0218 15:28:44.368130 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:28:44 crc kubenswrapper[4968]: I0218 15:28:44.368832 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:28:44 crc kubenswrapper[4968]: I0218 15:28:44.368958 4968 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:28:44 crc kubenswrapper[4968]: I0218 15:28:44.369996 4968 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5d6c8b54ab16252892873fa8a0af6e5bfc9fe64986de280b4035d7ef2a1e29f8"} pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:28:44 crc kubenswrapper[4968]: I0218 15:28:44.370161 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" containerID="cri-o://5d6c8b54ab16252892873fa8a0af6e5bfc9fe64986de280b4035d7ef2a1e29f8" gracePeriod=600 Feb 18 15:28:45 crc kubenswrapper[4968]: I0218 15:28:45.324323 4968 generic.go:334] "Generic (PLEG): container finished" podID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerID="5d6c8b54ab16252892873fa8a0af6e5bfc9fe64986de280b4035d7ef2a1e29f8" exitCode=0 Feb 18 15:28:45 crc kubenswrapper[4968]: I0218 15:28:45.324400 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerDied","Data":"5d6c8b54ab16252892873fa8a0af6e5bfc9fe64986de280b4035d7ef2a1e29f8"} Feb 18 15:28:45 crc kubenswrapper[4968]: I0218 15:28:45.325033 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerStarted","Data":"c21daacf9a62d32cfa427d33ebe792f5f91bea8de4e6e2738e79b36fa1f3f589"} Feb 18 15:28:45 crc kubenswrapper[4968]: I0218 15:28:45.325100 4968 scope.go:117] "RemoveContainer" containerID="940658e0b35b913490ed555675eeddf74061090b611d10aa557ed1bb4e8242b5" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.197861 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4"] Feb 18 15:30:00 crc kubenswrapper[4968]: E0218 15:30:00.200195 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" containerName="registry" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.200315 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" containerName="registry" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.200513 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="19f5e9c6-5b94-4c3a-8f01-3ed6aad36ec2" containerName="registry" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.201090 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.203546 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.204171 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.205125 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4"] Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.356558 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2sk2\" (UniqueName: \"kubernetes.io/projected/b8ff8bda-fbb2-4559-9416-10e196618fe9-kube-api-access-d2sk2\") pod \"collect-profiles-29523810-gsrn4\" (UID: \"b8ff8bda-fbb2-4559-9416-10e196618fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.356644 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8ff8bda-fbb2-4559-9416-10e196618fe9-secret-volume\") pod \"collect-profiles-29523810-gsrn4\" (UID: \"b8ff8bda-fbb2-4559-9416-10e196618fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.356698 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8ff8bda-fbb2-4559-9416-10e196618fe9-config-volume\") pod \"collect-profiles-29523810-gsrn4\" (UID: \"b8ff8bda-fbb2-4559-9416-10e196618fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.458590 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8ff8bda-fbb2-4559-9416-10e196618fe9-secret-volume\") pod \"collect-profiles-29523810-gsrn4\" (UID: \"b8ff8bda-fbb2-4559-9416-10e196618fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.458650 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8ff8bda-fbb2-4559-9416-10e196618fe9-config-volume\") pod \"collect-profiles-29523810-gsrn4\" (UID: \"b8ff8bda-fbb2-4559-9416-10e196618fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.458708 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2sk2\" (UniqueName: \"kubernetes.io/projected/b8ff8bda-fbb2-4559-9416-10e196618fe9-kube-api-access-d2sk2\") pod \"collect-profiles-29523810-gsrn4\" (UID: \"b8ff8bda-fbb2-4559-9416-10e196618fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.459800 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8ff8bda-fbb2-4559-9416-10e196618fe9-config-volume\") pod \"collect-profiles-29523810-gsrn4\" (UID: \"b8ff8bda-fbb2-4559-9416-10e196618fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.466990 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8ff8bda-fbb2-4559-9416-10e196618fe9-secret-volume\") pod \"collect-profiles-29523810-gsrn4\" (UID: \"b8ff8bda-fbb2-4559-9416-10e196618fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.482634 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2sk2\" (UniqueName: \"kubernetes.io/projected/b8ff8bda-fbb2-4559-9416-10e196618fe9-kube-api-access-d2sk2\") pod \"collect-profiles-29523810-gsrn4\" (UID: \"b8ff8bda-fbb2-4559-9416-10e196618fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.524359 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:00 crc kubenswrapper[4968]: I0218 15:30:00.712137 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4"] Feb 18 15:30:01 crc kubenswrapper[4968]: I0218 15:30:01.109764 4968 generic.go:334] "Generic (PLEG): container finished" podID="b8ff8bda-fbb2-4559-9416-10e196618fe9" containerID="642a4be6eb2b7d7023f70315c83a0d3c2b294529313eead4fc94a73848b7593d" exitCode=0 Feb 18 15:30:01 crc kubenswrapper[4968]: I0218 15:30:01.109810 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" event={"ID":"b8ff8bda-fbb2-4559-9416-10e196618fe9","Type":"ContainerDied","Data":"642a4be6eb2b7d7023f70315c83a0d3c2b294529313eead4fc94a73848b7593d"} Feb 18 15:30:01 crc kubenswrapper[4968]: I0218 15:30:01.109836 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" event={"ID":"b8ff8bda-fbb2-4559-9416-10e196618fe9","Type":"ContainerStarted","Data":"2c4c4018ddb94f22fb17ce68b45ae686cda8add44f0f65f133096de0a78d1414"} Feb 18 15:30:02 crc kubenswrapper[4968]: I0218 15:30:02.365420 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:02 crc kubenswrapper[4968]: I0218 15:30:02.483563 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8ff8bda-fbb2-4559-9416-10e196618fe9-config-volume\") pod \"b8ff8bda-fbb2-4559-9416-10e196618fe9\" (UID: \"b8ff8bda-fbb2-4559-9416-10e196618fe9\") " Feb 18 15:30:02 crc kubenswrapper[4968]: I0218 15:30:02.483622 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2sk2\" (UniqueName: \"kubernetes.io/projected/b8ff8bda-fbb2-4559-9416-10e196618fe9-kube-api-access-d2sk2\") pod \"b8ff8bda-fbb2-4559-9416-10e196618fe9\" (UID: \"b8ff8bda-fbb2-4559-9416-10e196618fe9\") " Feb 18 15:30:02 crc kubenswrapper[4968]: I0218 15:30:02.483682 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8ff8bda-fbb2-4559-9416-10e196618fe9-secret-volume\") pod \"b8ff8bda-fbb2-4559-9416-10e196618fe9\" (UID: \"b8ff8bda-fbb2-4559-9416-10e196618fe9\") " Feb 18 15:30:02 crc kubenswrapper[4968]: I0218 15:30:02.484610 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8ff8bda-fbb2-4559-9416-10e196618fe9-config-volume" (OuterVolumeSpecName: "config-volume") pod "b8ff8bda-fbb2-4559-9416-10e196618fe9" (UID: "b8ff8bda-fbb2-4559-9416-10e196618fe9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:30:02 crc kubenswrapper[4968]: I0218 15:30:02.488640 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ff8bda-fbb2-4559-9416-10e196618fe9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b8ff8bda-fbb2-4559-9416-10e196618fe9" (UID: "b8ff8bda-fbb2-4559-9416-10e196618fe9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:30:02 crc kubenswrapper[4968]: I0218 15:30:02.488895 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8ff8bda-fbb2-4559-9416-10e196618fe9-kube-api-access-d2sk2" (OuterVolumeSpecName: "kube-api-access-d2sk2") pod "b8ff8bda-fbb2-4559-9416-10e196618fe9" (UID: "b8ff8bda-fbb2-4559-9416-10e196618fe9"). InnerVolumeSpecName "kube-api-access-d2sk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:30:02 crc kubenswrapper[4968]: I0218 15:30:02.584895 4968 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8ff8bda-fbb2-4559-9416-10e196618fe9-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 18 15:30:02 crc kubenswrapper[4968]: I0218 15:30:02.584931 4968 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8ff8bda-fbb2-4559-9416-10e196618fe9-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 15:30:02 crc kubenswrapper[4968]: I0218 15:30:02.584945 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2sk2\" (UniqueName: \"kubernetes.io/projected/b8ff8bda-fbb2-4559-9416-10e196618fe9-kube-api-access-d2sk2\") on node \"crc\" DevicePath \"\"" Feb 18 15:30:03 crc kubenswrapper[4968]: I0218 15:30:03.124022 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" event={"ID":"b8ff8bda-fbb2-4559-9416-10e196618fe9","Type":"ContainerDied","Data":"2c4c4018ddb94f22fb17ce68b45ae686cda8add44f0f65f133096de0a78d1414"} Feb 18 15:30:03 crc kubenswrapper[4968]: I0218 15:30:03.124370 4968 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c4c4018ddb94f22fb17ce68b45ae686cda8add44f0f65f133096de0a78d1414" Feb 18 15:30:03 crc kubenswrapper[4968]: I0218 15:30:03.124097 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523810-gsrn4" Feb 18 15:30:44 crc kubenswrapper[4968]: I0218 15:30:44.367488 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:30:44 crc kubenswrapper[4968]: I0218 15:30:44.368378 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:31:14 crc kubenswrapper[4968]: I0218 15:31:14.367846 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:31:14 crc kubenswrapper[4968]: I0218 15:31:14.368560 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:31:44 crc kubenswrapper[4968]: I0218 15:31:44.367151 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:31:44 crc kubenswrapper[4968]: I0218 15:31:44.368328 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:31:44 crc kubenswrapper[4968]: I0218 15:31:44.368415 4968 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:31:44 crc kubenswrapper[4968]: I0218 15:31:44.369280 4968 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c21daacf9a62d32cfa427d33ebe792f5f91bea8de4e6e2738e79b36fa1f3f589"} pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:31:44 crc kubenswrapper[4968]: I0218 15:31:44.369385 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" containerID="cri-o://c21daacf9a62d32cfa427d33ebe792f5f91bea8de4e6e2738e79b36fa1f3f589" gracePeriod=600 Feb 18 15:31:44 crc kubenswrapper[4968]: I0218 15:31:44.784442 4968 generic.go:334] "Generic (PLEG): container finished" podID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerID="c21daacf9a62d32cfa427d33ebe792f5f91bea8de4e6e2738e79b36fa1f3f589" exitCode=0 Feb 18 15:31:44 crc kubenswrapper[4968]: I0218 15:31:44.784504 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerDied","Data":"c21daacf9a62d32cfa427d33ebe792f5f91bea8de4e6e2738e79b36fa1f3f589"} Feb 18 15:31:44 crc kubenswrapper[4968]: I0218 15:31:44.785004 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerStarted","Data":"47c8c50ad8111d46dcdc1997d2e8171ff7419555273a45fc7043517f33afc7b7"} Feb 18 15:31:44 crc kubenswrapper[4968]: I0218 15:31:44.785047 4968 scope.go:117] "RemoveContainer" containerID="5d6c8b54ab16252892873fa8a0af6e5bfc9fe64986de280b4035d7ef2a1e29f8" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.444638 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-x7mcq"] Feb 18 15:32:36 crc kubenswrapper[4968]: E0218 15:32:36.445684 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8ff8bda-fbb2-4559-9416-10e196618fe9" containerName="collect-profiles" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.445702 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8ff8bda-fbb2-4559-9416-10e196618fe9" containerName="collect-profiles" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.445904 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8ff8bda-fbb2-4559-9416-10e196618fe9" containerName="collect-profiles" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.446437 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-x7mcq" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.448575 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.448811 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.450062 4968 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-pwlh7" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.452896 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-98hkr"] Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.453804 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-98hkr" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.455666 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-wpkq2"] Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.456339 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-wpkq2" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.461578 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-x7mcq"] Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.463706 4968 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-59d2b" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.470172 4968 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-jsdt7" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.477190 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-98hkr"] Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.483931 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-wpkq2"] Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.560090 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfvwl\" (UniqueName: \"kubernetes.io/projected/f34e4121-0b18-48e0-b48e-54054ebfbc72-kube-api-access-jfvwl\") pod \"cert-manager-cainjector-cf98fcc89-x7mcq\" (UID: \"f34e4121-0b18-48e0-b48e-54054ebfbc72\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-x7mcq" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.560151 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsc65\" (UniqueName: \"kubernetes.io/projected/b2fbab5c-3c83-4630-bfba-48bd448b5449-kube-api-access-jsc65\") pod \"cert-manager-858654f9db-98hkr\" (UID: \"b2fbab5c-3c83-4630-bfba-48bd448b5449\") " pod="cert-manager/cert-manager-858654f9db-98hkr" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.560185 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9lrz\" (UniqueName: \"kubernetes.io/projected/30349f7d-e8d4-4791-8668-29e3213e3be3-kube-api-access-p9lrz\") pod \"cert-manager-webhook-687f57d79b-wpkq2\" (UID: \"30349f7d-e8d4-4791-8668-29e3213e3be3\") " pod="cert-manager/cert-manager-webhook-687f57d79b-wpkq2" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.662299 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfvwl\" (UniqueName: \"kubernetes.io/projected/f34e4121-0b18-48e0-b48e-54054ebfbc72-kube-api-access-jfvwl\") pod \"cert-manager-cainjector-cf98fcc89-x7mcq\" (UID: \"f34e4121-0b18-48e0-b48e-54054ebfbc72\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-x7mcq" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.662357 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsc65\" (UniqueName: \"kubernetes.io/projected/b2fbab5c-3c83-4630-bfba-48bd448b5449-kube-api-access-jsc65\") pod \"cert-manager-858654f9db-98hkr\" (UID: \"b2fbab5c-3c83-4630-bfba-48bd448b5449\") " pod="cert-manager/cert-manager-858654f9db-98hkr" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.662392 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9lrz\" (UniqueName: \"kubernetes.io/projected/30349f7d-e8d4-4791-8668-29e3213e3be3-kube-api-access-p9lrz\") pod \"cert-manager-webhook-687f57d79b-wpkq2\" (UID: \"30349f7d-e8d4-4791-8668-29e3213e3be3\") " pod="cert-manager/cert-manager-webhook-687f57d79b-wpkq2" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.694013 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsc65\" (UniqueName: \"kubernetes.io/projected/b2fbab5c-3c83-4630-bfba-48bd448b5449-kube-api-access-jsc65\") pod \"cert-manager-858654f9db-98hkr\" (UID: \"b2fbab5c-3c83-4630-bfba-48bd448b5449\") " pod="cert-manager/cert-manager-858654f9db-98hkr" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.695364 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfvwl\" (UniqueName: \"kubernetes.io/projected/f34e4121-0b18-48e0-b48e-54054ebfbc72-kube-api-access-jfvwl\") pod \"cert-manager-cainjector-cf98fcc89-x7mcq\" (UID: \"f34e4121-0b18-48e0-b48e-54054ebfbc72\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-x7mcq" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.698292 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9lrz\" (UniqueName: \"kubernetes.io/projected/30349f7d-e8d4-4791-8668-29e3213e3be3-kube-api-access-p9lrz\") pod \"cert-manager-webhook-687f57d79b-wpkq2\" (UID: \"30349f7d-e8d4-4791-8668-29e3213e3be3\") " pod="cert-manager/cert-manager-webhook-687f57d79b-wpkq2" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.786550 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-x7mcq" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.795015 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-98hkr" Feb 18 15:32:36 crc kubenswrapper[4968]: I0218 15:32:36.802250 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-wpkq2" Feb 18 15:32:37 crc kubenswrapper[4968]: I0218 15:32:37.134301 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-wpkq2"] Feb 18 15:32:37 crc kubenswrapper[4968]: I0218 15:32:37.146078 4968 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 15:32:37 crc kubenswrapper[4968]: I0218 15:32:37.186884 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-98hkr"] Feb 18 15:32:37 crc kubenswrapper[4968]: W0218 15:32:37.193730 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2fbab5c_3c83_4630_bfba_48bd448b5449.slice/crio-c1bef6f419d222f4277ff2389a43714ca312864e8d02495b736505b6be467a88 WatchSource:0}: Error finding container c1bef6f419d222f4277ff2389a43714ca312864e8d02495b736505b6be467a88: Status 404 returned error can't find the container with id c1bef6f419d222f4277ff2389a43714ca312864e8d02495b736505b6be467a88 Feb 18 15:32:37 crc kubenswrapper[4968]: I0218 15:32:37.213794 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-x7mcq"] Feb 18 15:32:37 crc kubenswrapper[4968]: W0218 15:32:37.217060 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf34e4121_0b18_48e0_b48e_54054ebfbc72.slice/crio-6d25d90f6e9c663d03cc4dc1af2d163812dc9d13cf103787fb08d417fd0ed74e WatchSource:0}: Error finding container 6d25d90f6e9c663d03cc4dc1af2d163812dc9d13cf103787fb08d417fd0ed74e: Status 404 returned error can't find the container with id 6d25d90f6e9c663d03cc4dc1af2d163812dc9d13cf103787fb08d417fd0ed74e Feb 18 15:32:38 crc kubenswrapper[4968]: I0218 15:32:38.105542 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-98hkr" event={"ID":"b2fbab5c-3c83-4630-bfba-48bd448b5449","Type":"ContainerStarted","Data":"c1bef6f419d222f4277ff2389a43714ca312864e8d02495b736505b6be467a88"} Feb 18 15:32:38 crc kubenswrapper[4968]: I0218 15:32:38.107028 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-x7mcq" event={"ID":"f34e4121-0b18-48e0-b48e-54054ebfbc72","Type":"ContainerStarted","Data":"6d25d90f6e9c663d03cc4dc1af2d163812dc9d13cf103787fb08d417fd0ed74e"} Feb 18 15:32:38 crc kubenswrapper[4968]: I0218 15:32:38.109166 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-wpkq2" event={"ID":"30349f7d-e8d4-4791-8668-29e3213e3be3","Type":"ContainerStarted","Data":"5924b5c52a923462f83f2d561d026eed3b73b76a7a43a61c0ebf3b539d2c4e0c"} Feb 18 15:32:41 crc kubenswrapper[4968]: I0218 15:32:41.124719 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-98hkr" event={"ID":"b2fbab5c-3c83-4630-bfba-48bd448b5449","Type":"ContainerStarted","Data":"12d33f9bb163d7846c7eb1f569158697b963670ac7c45ca26e8d1bf0912a1f1e"} Feb 18 15:32:41 crc kubenswrapper[4968]: I0218 15:32:41.126959 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-x7mcq" event={"ID":"f34e4121-0b18-48e0-b48e-54054ebfbc72","Type":"ContainerStarted","Data":"39c613985dda0986ae3946c13cdf331a430e79ba73c05a522ff84ad83799c069"} Feb 18 15:32:41 crc kubenswrapper[4968]: I0218 15:32:41.128210 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-wpkq2" event={"ID":"30349f7d-e8d4-4791-8668-29e3213e3be3","Type":"ContainerStarted","Data":"dd7283e8a86056496c2c56aab5c2dbf181dc725b30c48df3c3b02d11ba18cf85"} Feb 18 15:32:41 crc kubenswrapper[4968]: I0218 15:32:41.128581 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-wpkq2" Feb 18 15:32:41 crc kubenswrapper[4968]: I0218 15:32:41.140672 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-98hkr" podStartSLOduration=1.7420234589999999 podStartE2EDuration="5.140655049s" podCreationTimestamp="2026-02-18 15:32:36 +0000 UTC" firstStartedPulling="2026-02-18 15:32:37.195788088 +0000 UTC m=+656.581232960" lastFinishedPulling="2026-02-18 15:32:40.594419688 +0000 UTC m=+659.979864550" observedRunningTime="2026-02-18 15:32:41.140010901 +0000 UTC m=+660.525455763" watchObservedRunningTime="2026-02-18 15:32:41.140655049 +0000 UTC m=+660.526099911" Feb 18 15:32:41 crc kubenswrapper[4968]: I0218 15:32:41.165239 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-wpkq2" podStartSLOduration=1.793047777 podStartE2EDuration="5.165198886s" podCreationTimestamp="2026-02-18 15:32:36 +0000 UTC" firstStartedPulling="2026-02-18 15:32:37.145850941 +0000 UTC m=+656.531295793" lastFinishedPulling="2026-02-18 15:32:40.51800205 +0000 UTC m=+659.903446902" observedRunningTime="2026-02-18 15:32:41.161497672 +0000 UTC m=+660.546942534" watchObservedRunningTime="2026-02-18 15:32:41.165198886 +0000 UTC m=+660.550643748" Feb 18 15:32:41 crc kubenswrapper[4968]: I0218 15:32:41.176622 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-x7mcq" podStartSLOduration=1.878030384 podStartE2EDuration="5.176602745s" podCreationTimestamp="2026-02-18 15:32:36 +0000 UTC" firstStartedPulling="2026-02-18 15:32:37.219607684 +0000 UTC m=+656.605052546" lastFinishedPulling="2026-02-18 15:32:40.518180035 +0000 UTC m=+659.903624907" observedRunningTime="2026-02-18 15:32:41.175823713 +0000 UTC m=+660.561268585" watchObservedRunningTime="2026-02-18 15:32:41.176602745 +0000 UTC m=+660.562047607" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.498583 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-z2jkk"] Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.500395 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovn-controller" containerID="cri-o://bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f" gracePeriod=30 Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.500814 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c" gracePeriod=30 Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.500837 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="nbdb" containerID="cri-o://1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608" gracePeriod=30 Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.500894 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="northd" containerID="cri-o://76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c" gracePeriod=30 Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.500910 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="sbdb" containerID="cri-o://29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49" gracePeriod=30 Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.500935 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovn-acl-logging" containerID="cri-o://04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2" gracePeriod=30 Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.501111 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="kube-rbac-proxy-node" containerID="cri-o://bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a" gracePeriod=30 Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.532999 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" containerID="cri-o://555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b" gracePeriod=30 Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.804736 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-wpkq2" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.874905 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/3.log" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.877587 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovn-acl-logging/0.log" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.878146 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovn-controller/0.log" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.878918 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945256 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-g7ffz"] Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.945503 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="northd" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945518 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="northd" Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.945535 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="kube-rbac-proxy-node" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945545 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="kube-rbac-proxy-node" Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.945555 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945564 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.945577 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="kubecfg-setup" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945585 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="kubecfg-setup" Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.945598 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945606 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.945615 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovn-acl-logging" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945625 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovn-acl-logging" Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.945636 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945644 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.945659 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945668 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.945679 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="sbdb" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945686 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="sbdb" Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.945698 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="nbdb" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945706 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="nbdb" Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.945715 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="kube-rbac-proxy-ovn-metrics" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945723 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="kube-rbac-proxy-ovn-metrics" Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.945732 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovn-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945739 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovn-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945889 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="nbdb" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945906 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="sbdb" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945916 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovn-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945926 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovn-acl-logging" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945936 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945945 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945955 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="northd" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945974 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945983 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="kube-rbac-proxy-node" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.945992 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="kube-rbac-proxy-ovn-metrics" Feb 18 15:32:46 crc kubenswrapper[4968]: E0218 15:32:46.946108 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.946119 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.946242 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.946255 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerName="ovnkube-controller" Feb 18 15:32:46 crc kubenswrapper[4968]: I0218 15:32:46.948186 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.040093 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-run-netns\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.040523 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovnkube-config\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.040571 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-env-overrides\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.040618 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hmgv\" (UniqueName: \"kubernetes.io/projected/435c6f94-b91e-4ce0-8407-5227f3a5078f-kube-api-access-5hmgv\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.040654 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-systemd-units\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.040680 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-etc-openvswitch\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.040720 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-ovn\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.040787 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovn-node-metrics-cert\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.040833 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-log-socket\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.040875 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-openvswitch\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.040259 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.040964 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041020 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041106 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-log-socket" (OuterVolumeSpecName: "log-socket") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041107 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041203 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041397 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041460 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041529 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovnkube-script-lib\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041625 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-slash\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041663 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-slash" (OuterVolumeSpecName: "host-slash") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041844 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-cni-netd\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041892 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-node-log\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041895 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.041921 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-node-log" (OuterVolumeSpecName: "node-log") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042042 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-cni-bin\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042066 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-systemd\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042086 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042107 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-kubelet\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042144 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-run-ovn-kubernetes\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042172 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-var-lib-openvswitch\") pod \"435c6f94-b91e-4ce0-8407-5227f3a5078f\" (UID: \"435c6f94-b91e-4ce0-8407-5227f3a5078f\") " Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042194 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042210 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042227 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042236 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042337 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-var-lib-openvswitch\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042348 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042365 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-run-ovn-kubernetes\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042506 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-log-socket\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042535 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042552 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-run-systemd\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042571 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-ovn-node-metrics-cert\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042608 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-env-overrides\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042632 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-cni-bin\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042631 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042655 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-etc-openvswitch\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042676 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-run-openvswitch\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042698 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-slash\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042780 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-ovnkube-config\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042839 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-kubelet\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042904 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-systemd-units\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.042961 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-ovnkube-script-lib\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043025 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-cni-netd\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043061 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-node-log\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043137 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9qxb\" (UniqueName: \"kubernetes.io/projected/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-kube-api-access-h9qxb\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043175 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-run-ovn\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043199 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-run-netns\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043246 4968 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043255 4968 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-log-socket\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043264 4968 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043273 4968 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043281 4968 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-slash\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043289 4968 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043322 4968 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-node-log\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043352 4968 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043374 4968 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043392 4968 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043410 4968 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043426 4968 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043444 4968 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043459 4968 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043474 4968 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/435c6f94-b91e-4ce0-8407-5227f3a5078f-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043488 4968 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.043503 4968 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.047921 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/435c6f94-b91e-4ce0-8407-5227f3a5078f-kube-api-access-5hmgv" (OuterVolumeSpecName: "kube-api-access-5hmgv") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "kube-api-access-5hmgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.047929 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.056174 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "435c6f94-b91e-4ce0-8407-5227f3a5078f" (UID: "435c6f94-b91e-4ce0-8407-5227f3a5078f"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.143835 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-run-netns\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.143891 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-var-lib-openvswitch\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.143915 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-run-ovn-kubernetes\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.143948 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-log-socket\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.143975 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.143986 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-run-ovn-kubernetes\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144033 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-run-systemd\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.143994 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-run-systemd\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144042 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-log-socket\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.143940 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-run-netns\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144074 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144066 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-var-lib-openvswitch\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144100 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-ovn-node-metrics-cert\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144211 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-env-overrides\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144250 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-cni-bin\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144292 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-etc-openvswitch\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144323 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-run-openvswitch\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144386 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-slash\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144462 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-ovnkube-config\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144522 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-kubelet\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144578 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-systemd-units\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144623 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-ovnkube-script-lib\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144672 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-cni-netd\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144712 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-node-log\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144809 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9qxb\" (UniqueName: \"kubernetes.io/projected/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-kube-api-access-h9qxb\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144846 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-etc-openvswitch\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144858 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-run-ovn\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144891 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-cni-bin\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144914 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-node-log\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144951 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-cni-netd\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144953 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-run-ovn\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144971 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-kubelet\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.144997 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-env-overrides\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.145025 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-systemd-units\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.145026 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-run-openvswitch\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.145092 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-host-slash\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.145158 4968 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/435c6f94-b91e-4ce0-8407-5227f3a5078f-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.145178 4968 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/435c6f94-b91e-4ce0-8407-5227f3a5078f-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.145223 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hmgv\" (UniqueName: \"kubernetes.io/projected/435c6f94-b91e-4ce0-8407-5227f3a5078f-kube-api-access-5hmgv\") on node \"crc\" DevicePath \"\"" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.145704 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-ovnkube-script-lib\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.145859 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-ovnkube-config\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.147907 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-ovn-node-metrics-cert\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.161596 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9qxb\" (UniqueName: \"kubernetes.io/projected/ef51e88e-fae0-4037-97e6-9a598dcc4a3f-kube-api-access-h9qxb\") pod \"ovnkube-node-g7ffz\" (UID: \"ef51e88e-fae0-4037-97e6-9a598dcc4a3f\") " pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.165415 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2qq8_2d01781a-6a78-49a2-80c7-9ac02c810e3f/kube-multus/2.log" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.165918 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2qq8_2d01781a-6a78-49a2-80c7-9ac02c810e3f/kube-multus/1.log" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.165957 4968 generic.go:334] "Generic (PLEG): container finished" podID="2d01781a-6a78-49a2-80c7-9ac02c810e3f" containerID="46fd3a8731ea670c4e3883ecc9aa92f701aedd645c1658c7f936d0005d273013" exitCode=2 Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.166009 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2qq8" event={"ID":"2d01781a-6a78-49a2-80c7-9ac02c810e3f","Type":"ContainerDied","Data":"46fd3a8731ea670c4e3883ecc9aa92f701aedd645c1658c7f936d0005d273013"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.166045 4968 scope.go:117] "RemoveContainer" containerID="b6ccc51a3ebdf422ba17225bac5935ff711e8ed81619d01a461692c3fb9083eb" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.166566 4968 scope.go:117] "RemoveContainer" containerID="46fd3a8731ea670c4e3883ecc9aa92f701aedd645c1658c7f936d0005d273013" Feb 18 15:32:47 crc kubenswrapper[4968]: E0218 15:32:47.166820 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-m2qq8_openshift-multus(2d01781a-6a78-49a2-80c7-9ac02c810e3f)\"" pod="openshift-multus/multus-m2qq8" podUID="2d01781a-6a78-49a2-80c7-9ac02c810e3f" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.170277 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovnkube-controller/3.log" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.172921 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovn-acl-logging/0.log" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.173526 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z2jkk_435c6f94-b91e-4ce0-8407-5227f3a5078f/ovn-controller/0.log" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174088 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b" exitCode=0 Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174137 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49" exitCode=0 Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174162 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608" exitCode=0 Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174183 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c" exitCode=0 Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174187 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174227 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174247 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174262 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174277 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174199 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c" exitCode=0 Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174305 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a" exitCode=0 Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174334 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2" exitCode=143 Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174345 4968 generic.go:334] "Generic (PLEG): container finished" podID="435c6f94-b91e-4ce0-8407-5227f3a5078f" containerID="bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f" exitCode=143 Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174361 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174375 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174388 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174396 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174404 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174411 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174446 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174457 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174464 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174472 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174479 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174490 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174502 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174510 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174518 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174525 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174532 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174540 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174547 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174554 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174561 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174568 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174578 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174589 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174597 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174605 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174613 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174620 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174627 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174635 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174643 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174651 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174658 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174668 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" event={"ID":"435c6f94-b91e-4ce0-8407-5227f3a5078f","Type":"ContainerDied","Data":"a5a166ddc4cb02e8be575a1382f44a5f44030729235c4926c78175404505570f"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174679 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174687 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174694 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174702 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174708 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174715 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174722 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174729 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174736 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.174763 4968 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502"} Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.175708 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-z2jkk" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.207106 4968 scope.go:117] "RemoveContainer" containerID="555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.226897 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-z2jkk"] Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.238039 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-z2jkk"] Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.241141 4968 scope.go:117] "RemoveContainer" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.258580 4968 scope.go:117] "RemoveContainer" containerID="29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.266970 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.273315 4968 scope.go:117] "RemoveContainer" containerID="1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.293282 4968 scope.go:117] "RemoveContainer" containerID="76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.308768 4968 scope.go:117] "RemoveContainer" containerID="727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.327335 4968 scope.go:117] "RemoveContainer" containerID="bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.343309 4968 scope.go:117] "RemoveContainer" containerID="04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.367211 4968 scope.go:117] "RemoveContainer" containerID="bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.406347 4968 scope.go:117] "RemoveContainer" containerID="0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.427718 4968 scope.go:117] "RemoveContainer" containerID="555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b" Feb 18 15:32:47 crc kubenswrapper[4968]: E0218 15:32:47.428165 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b\": container with ID starting with 555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b not found: ID does not exist" containerID="555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.428198 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b"} err="failed to get container status \"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b\": rpc error: code = NotFound desc = could not find container \"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b\": container with ID starting with 555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.428224 4968 scope.go:117] "RemoveContainer" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" Feb 18 15:32:47 crc kubenswrapper[4968]: E0218 15:32:47.428974 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\": container with ID starting with 945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3 not found: ID does not exist" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.428995 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3"} err="failed to get container status \"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\": rpc error: code = NotFound desc = could not find container \"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\": container with ID starting with 945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.429013 4968 scope.go:117] "RemoveContainer" containerID="29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49" Feb 18 15:32:47 crc kubenswrapper[4968]: E0218 15:32:47.429356 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\": container with ID starting with 29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49 not found: ID does not exist" containerID="29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.429376 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49"} err="failed to get container status \"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\": rpc error: code = NotFound desc = could not find container \"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\": container with ID starting with 29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.429392 4968 scope.go:117] "RemoveContainer" containerID="1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608" Feb 18 15:32:47 crc kubenswrapper[4968]: E0218 15:32:47.429947 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\": container with ID starting with 1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608 not found: ID does not exist" containerID="1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.429968 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608"} err="failed to get container status \"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\": rpc error: code = NotFound desc = could not find container \"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\": container with ID starting with 1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.429985 4968 scope.go:117] "RemoveContainer" containerID="76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c" Feb 18 15:32:47 crc kubenswrapper[4968]: E0218 15:32:47.430521 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\": container with ID starting with 76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c not found: ID does not exist" containerID="76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.430545 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c"} err="failed to get container status \"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\": rpc error: code = NotFound desc = could not find container \"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\": container with ID starting with 76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.430562 4968 scope.go:117] "RemoveContainer" containerID="727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c" Feb 18 15:32:47 crc kubenswrapper[4968]: E0218 15:32:47.430979 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\": container with ID starting with 727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c not found: ID does not exist" containerID="727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.431063 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c"} err="failed to get container status \"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\": rpc error: code = NotFound desc = could not find container \"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\": container with ID starting with 727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.431129 4968 scope.go:117] "RemoveContainer" containerID="bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a" Feb 18 15:32:47 crc kubenswrapper[4968]: E0218 15:32:47.431540 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\": container with ID starting with bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a not found: ID does not exist" containerID="bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.431564 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a"} err="failed to get container status \"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\": rpc error: code = NotFound desc = could not find container \"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\": container with ID starting with bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.431586 4968 scope.go:117] "RemoveContainer" containerID="04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2" Feb 18 15:32:47 crc kubenswrapper[4968]: E0218 15:32:47.432134 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\": container with ID starting with 04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2 not found: ID does not exist" containerID="04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.432154 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2"} err="failed to get container status \"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\": rpc error: code = NotFound desc = could not find container \"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\": container with ID starting with 04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.432172 4968 scope.go:117] "RemoveContainer" containerID="bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f" Feb 18 15:32:47 crc kubenswrapper[4968]: E0218 15:32:47.432466 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\": container with ID starting with bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f not found: ID does not exist" containerID="bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.432509 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f"} err="failed to get container status \"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\": rpc error: code = NotFound desc = could not find container \"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\": container with ID starting with bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.432541 4968 scope.go:117] "RemoveContainer" containerID="0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502" Feb 18 15:32:47 crc kubenswrapper[4968]: E0218 15:32:47.433051 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\": container with ID starting with 0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502 not found: ID does not exist" containerID="0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.433118 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502"} err="failed to get container status \"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\": rpc error: code = NotFound desc = could not find container \"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\": container with ID starting with 0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.433158 4968 scope.go:117] "RemoveContainer" containerID="555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.433697 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b"} err="failed to get container status \"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b\": rpc error: code = NotFound desc = could not find container \"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b\": container with ID starting with 555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.433781 4968 scope.go:117] "RemoveContainer" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.434291 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3"} err="failed to get container status \"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\": rpc error: code = NotFound desc = could not find container \"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\": container with ID starting with 945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.434328 4968 scope.go:117] "RemoveContainer" containerID="29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.434698 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49"} err="failed to get container status \"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\": rpc error: code = NotFound desc = could not find container \"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\": container with ID starting with 29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.434730 4968 scope.go:117] "RemoveContainer" containerID="1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.435163 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608"} err="failed to get container status \"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\": rpc error: code = NotFound desc = could not find container \"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\": container with ID starting with 1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.435205 4968 scope.go:117] "RemoveContainer" containerID="76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.435593 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c"} err="failed to get container status \"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\": rpc error: code = NotFound desc = could not find container \"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\": container with ID starting with 76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.435624 4968 scope.go:117] "RemoveContainer" containerID="727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.436430 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c"} err="failed to get container status \"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\": rpc error: code = NotFound desc = could not find container \"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\": container with ID starting with 727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.436486 4968 scope.go:117] "RemoveContainer" containerID="bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.436892 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a"} err="failed to get container status \"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\": rpc error: code = NotFound desc = could not find container \"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\": container with ID starting with bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.436924 4968 scope.go:117] "RemoveContainer" containerID="04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.437283 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2"} err="failed to get container status \"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\": rpc error: code = NotFound desc = could not find container \"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\": container with ID starting with 04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.437318 4968 scope.go:117] "RemoveContainer" containerID="bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.437684 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f"} err="failed to get container status \"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\": rpc error: code = NotFound desc = could not find container \"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\": container with ID starting with bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.437776 4968 scope.go:117] "RemoveContainer" containerID="0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.438245 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502"} err="failed to get container status \"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\": rpc error: code = NotFound desc = could not find container \"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\": container with ID starting with 0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.438272 4968 scope.go:117] "RemoveContainer" containerID="555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.438566 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b"} err="failed to get container status \"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b\": rpc error: code = NotFound desc = could not find container \"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b\": container with ID starting with 555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.438590 4968 scope.go:117] "RemoveContainer" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.438903 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3"} err="failed to get container status \"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\": rpc error: code = NotFound desc = could not find container \"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\": container with ID starting with 945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.438923 4968 scope.go:117] "RemoveContainer" containerID="29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.439302 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49"} err="failed to get container status \"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\": rpc error: code = NotFound desc = could not find container \"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\": container with ID starting with 29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.439321 4968 scope.go:117] "RemoveContainer" containerID="1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.439695 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608"} err="failed to get container status \"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\": rpc error: code = NotFound desc = could not find container \"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\": container with ID starting with 1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.439715 4968 scope.go:117] "RemoveContainer" containerID="76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.440063 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c"} err="failed to get container status \"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\": rpc error: code = NotFound desc = could not find container \"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\": container with ID starting with 76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.440099 4968 scope.go:117] "RemoveContainer" containerID="727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.440931 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c"} err="failed to get container status \"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\": rpc error: code = NotFound desc = could not find container \"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\": container with ID starting with 727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.441000 4968 scope.go:117] "RemoveContainer" containerID="bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.441432 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a"} err="failed to get container status \"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\": rpc error: code = NotFound desc = could not find container \"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\": container with ID starting with bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.441465 4968 scope.go:117] "RemoveContainer" containerID="04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.441874 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2"} err="failed to get container status \"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\": rpc error: code = NotFound desc = could not find container \"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\": container with ID starting with 04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.441937 4968 scope.go:117] "RemoveContainer" containerID="bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.442529 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f"} err="failed to get container status \"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\": rpc error: code = NotFound desc = could not find container \"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\": container with ID starting with bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.442564 4968 scope.go:117] "RemoveContainer" containerID="0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.443044 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502"} err="failed to get container status \"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\": rpc error: code = NotFound desc = could not find container \"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\": container with ID starting with 0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.443081 4968 scope.go:117] "RemoveContainer" containerID="555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.443436 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b"} err="failed to get container status \"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b\": rpc error: code = NotFound desc = could not find container \"555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b\": container with ID starting with 555c5d33aa78135e9cf0c7adff8fa8f94737d4be696dba5f11397f7bc8243c1b not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.443469 4968 scope.go:117] "RemoveContainer" containerID="945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.444057 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3"} err="failed to get container status \"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\": rpc error: code = NotFound desc = could not find container \"945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3\": container with ID starting with 945bc2e75017ac7a721f164a65fce369e108f791ed1289717f7e14b6ef2c5cb3 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.444087 4968 scope.go:117] "RemoveContainer" containerID="29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.444523 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49"} err="failed to get container status \"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\": rpc error: code = NotFound desc = could not find container \"29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49\": container with ID starting with 29b1e5e827bdf9f098441e38663290a253c9d2a54130630c2d0f9209ecef9c49 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.444559 4968 scope.go:117] "RemoveContainer" containerID="1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.444891 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608"} err="failed to get container status \"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\": rpc error: code = NotFound desc = could not find container \"1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608\": container with ID starting with 1b4d59e5845f202cee63202bdceb62267b2b196876b822565375f09b2ff7c608 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.444951 4968 scope.go:117] "RemoveContainer" containerID="76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.445359 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c"} err="failed to get container status \"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\": rpc error: code = NotFound desc = could not find container \"76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c\": container with ID starting with 76e676b0e106d3c008e5de61c6b110f6041737155997586ed6f13293c852539c not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.445380 4968 scope.go:117] "RemoveContainer" containerID="727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.445913 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c"} err="failed to get container status \"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\": rpc error: code = NotFound desc = could not find container \"727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c\": container with ID starting with 727a4b77091552280b1284ad912e79222795a47cb5dcdb2e0877705169e3853c not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.445935 4968 scope.go:117] "RemoveContainer" containerID="bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.446315 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a"} err="failed to get container status \"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\": rpc error: code = NotFound desc = could not find container \"bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a\": container with ID starting with bab43ae15f989e13a69319aaadd29ab217c393c8011705edcc8e96f42e294a8a not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.446333 4968 scope.go:117] "RemoveContainer" containerID="04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.447298 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2"} err="failed to get container status \"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\": rpc error: code = NotFound desc = could not find container \"04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2\": container with ID starting with 04f2fac0d793ec356290986138adffac24499ac1c9f4346a74fe3585b3bcc1f2 not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.448529 4968 scope.go:117] "RemoveContainer" containerID="bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.449003 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f"} err="failed to get container status \"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\": rpc error: code = NotFound desc = could not find container \"bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f\": container with ID starting with bec09fd048cd00a8afc336a344d7206460799d02bf44aec7f15d14999798da3f not found: ID does not exist" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.449034 4968 scope.go:117] "RemoveContainer" containerID="0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502" Feb 18 15:32:47 crc kubenswrapper[4968]: I0218 15:32:47.449539 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502"} err="failed to get container status \"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\": rpc error: code = NotFound desc = could not find container \"0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502\": container with ID starting with 0730631bc3b47f4b6ddf4bc4c3ebc9ff2290d5ee52814788d9ad46bc768e9502 not found: ID does not exist" Feb 18 15:32:48 crc kubenswrapper[4968]: I0218 15:32:48.182813 4968 generic.go:334] "Generic (PLEG): container finished" podID="ef51e88e-fae0-4037-97e6-9a598dcc4a3f" containerID="94df0ca3fd0998e4289f57d609eac29bf57489e1ee33fa15f1fb59e6c9c92f35" exitCode=0 Feb 18 15:32:48 crc kubenswrapper[4968]: I0218 15:32:48.182893 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" event={"ID":"ef51e88e-fae0-4037-97e6-9a598dcc4a3f","Type":"ContainerDied","Data":"94df0ca3fd0998e4289f57d609eac29bf57489e1ee33fa15f1fb59e6c9c92f35"} Feb 18 15:32:48 crc kubenswrapper[4968]: I0218 15:32:48.182965 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" event={"ID":"ef51e88e-fae0-4037-97e6-9a598dcc4a3f","Type":"ContainerStarted","Data":"079d170e14d6fa25429df5f2d78282f80151210f7ffae964a508ce57be3d4ecc"} Feb 18 15:32:48 crc kubenswrapper[4968]: I0218 15:32:48.185611 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2qq8_2d01781a-6a78-49a2-80c7-9ac02c810e3f/kube-multus/2.log" Feb 18 15:32:49 crc kubenswrapper[4968]: I0218 15:32:49.200995 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" event={"ID":"ef51e88e-fae0-4037-97e6-9a598dcc4a3f","Type":"ContainerStarted","Data":"b42e0964e6baec2a5b5cd3fb8d4da63176ea69ad88cb2dfdeb10b7f6c3a455c0"} Feb 18 15:32:49 crc kubenswrapper[4968]: I0218 15:32:49.201994 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" event={"ID":"ef51e88e-fae0-4037-97e6-9a598dcc4a3f","Type":"ContainerStarted","Data":"ebbeaf20784b241cbee24e8b02fcc1d69a7916e7b9d64a1a5e7f9d8c66c5fd94"} Feb 18 15:32:49 crc kubenswrapper[4968]: I0218 15:32:49.202049 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" event={"ID":"ef51e88e-fae0-4037-97e6-9a598dcc4a3f","Type":"ContainerStarted","Data":"322d0f078fbb27d859d54086c11c037400f3de2c2e2c7ec85cba13786ad9575d"} Feb 18 15:32:49 crc kubenswrapper[4968]: I0218 15:32:49.202069 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" event={"ID":"ef51e88e-fae0-4037-97e6-9a598dcc4a3f","Type":"ContainerStarted","Data":"91b957e221b81e875745037e6e58f72451c7c5c8a64fffc8890c895561952eda"} Feb 18 15:32:49 crc kubenswrapper[4968]: I0218 15:32:49.202091 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" event={"ID":"ef51e88e-fae0-4037-97e6-9a598dcc4a3f","Type":"ContainerStarted","Data":"d7ce7c39ccc30f875a74217170140648126d2ad3df7d3136a2493aececb4232d"} Feb 18 15:32:49 crc kubenswrapper[4968]: I0218 15:32:49.202114 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" event={"ID":"ef51e88e-fae0-4037-97e6-9a598dcc4a3f","Type":"ContainerStarted","Data":"cc86c32f888bf45c969118a81d4ed411068c34d5ac55655db56651aa913d0048"} Feb 18 15:32:49 crc kubenswrapper[4968]: I0218 15:32:49.243347 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="435c6f94-b91e-4ce0-8407-5227f3a5078f" path="/var/lib/kubelet/pods/435c6f94-b91e-4ce0-8407-5227f3a5078f/volumes" Feb 18 15:32:51 crc kubenswrapper[4968]: I0218 15:32:51.217237 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" event={"ID":"ef51e88e-fae0-4037-97e6-9a598dcc4a3f","Type":"ContainerStarted","Data":"e44e079f9255ff0584a5d1939f7f8e1f1e3a3e1d9faf10cbf71545dbbb838b4a"} Feb 18 15:32:54 crc kubenswrapper[4968]: I0218 15:32:54.242607 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" event={"ID":"ef51e88e-fae0-4037-97e6-9a598dcc4a3f","Type":"ContainerStarted","Data":"f30ef6f8facd7e000b1ba2888d2809bfdd6f1b6b0abb1671976382631df3b915"} Feb 18 15:32:54 crc kubenswrapper[4968]: I0218 15:32:54.243173 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:54 crc kubenswrapper[4968]: I0218 15:32:54.277380 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:54 crc kubenswrapper[4968]: I0218 15:32:54.290909 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" podStartSLOduration=8.29089298 podStartE2EDuration="8.29089298s" podCreationTimestamp="2026-02-18 15:32:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:32:54.289075929 +0000 UTC m=+673.674520811" watchObservedRunningTime="2026-02-18 15:32:54.29089298 +0000 UTC m=+673.676337842" Feb 18 15:32:55 crc kubenswrapper[4968]: I0218 15:32:55.257389 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:55 crc kubenswrapper[4968]: I0218 15:32:55.257443 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:55 crc kubenswrapper[4968]: I0218 15:32:55.302905 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:32:58 crc kubenswrapper[4968]: I0218 15:32:58.231548 4968 scope.go:117] "RemoveContainer" containerID="46fd3a8731ea670c4e3883ecc9aa92f701aedd645c1658c7f936d0005d273013" Feb 18 15:32:58 crc kubenswrapper[4968]: E0218 15:32:58.232265 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-m2qq8_openshift-multus(2d01781a-6a78-49a2-80c7-9ac02c810e3f)\"" pod="openshift-multus/multus-m2qq8" podUID="2d01781a-6a78-49a2-80c7-9ac02c810e3f" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.230917 4968 scope.go:117] "RemoveContainer" containerID="46fd3a8731ea670c4e3883ecc9aa92f701aedd645c1658c7f936d0005d273013" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.367145 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8"] Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.368675 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.371686 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.407462 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8"] Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.410620 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/62b19fba-8472-4001-b221-2d4f0dd1d36e-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8\" (UID: \"62b19fba-8472-4001-b221-2d4f0dd1d36e\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.410861 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cggc\" (UniqueName: \"kubernetes.io/projected/62b19fba-8472-4001-b221-2d4f0dd1d36e-kube-api-access-4cggc\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8\" (UID: \"62b19fba-8472-4001-b221-2d4f0dd1d36e\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.410941 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/62b19fba-8472-4001-b221-2d4f0dd1d36e-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8\" (UID: \"62b19fba-8472-4001-b221-2d4f0dd1d36e\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.511915 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/62b19fba-8472-4001-b221-2d4f0dd1d36e-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8\" (UID: \"62b19fba-8472-4001-b221-2d4f0dd1d36e\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.511979 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/62b19fba-8472-4001-b221-2d4f0dd1d36e-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8\" (UID: \"62b19fba-8472-4001-b221-2d4f0dd1d36e\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.512044 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cggc\" (UniqueName: \"kubernetes.io/projected/62b19fba-8472-4001-b221-2d4f0dd1d36e-kube-api-access-4cggc\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8\" (UID: \"62b19fba-8472-4001-b221-2d4f0dd1d36e\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.512523 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/62b19fba-8472-4001-b221-2d4f0dd1d36e-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8\" (UID: \"62b19fba-8472-4001-b221-2d4f0dd1d36e\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.512562 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/62b19fba-8472-4001-b221-2d4f0dd1d36e-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8\" (UID: \"62b19fba-8472-4001-b221-2d4f0dd1d36e\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.561128 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cggc\" (UniqueName: \"kubernetes.io/projected/62b19fba-8472-4001-b221-2d4f0dd1d36e-kube-api-access-4cggc\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8\" (UID: \"62b19fba-8472-4001-b221-2d4f0dd1d36e\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: I0218 15:33:13.698439 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: E0218 15:33:13.729732 4968 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_openshift-marketplace_62b19fba-8472-4001-b221-2d4f0dd1d36e_0(472cb6366899081d28bf55d0c34514db2bfe0409ca2da653ec1ed779f4c8ee62): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 18 15:33:13 crc kubenswrapper[4968]: E0218 15:33:13.729886 4968 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_openshift-marketplace_62b19fba-8472-4001-b221-2d4f0dd1d36e_0(472cb6366899081d28bf55d0c34514db2bfe0409ca2da653ec1ed779f4c8ee62): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: E0218 15:33:13.729938 4968 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_openshift-marketplace_62b19fba-8472-4001-b221-2d4f0dd1d36e_0(472cb6366899081d28bf55d0c34514db2bfe0409ca2da653ec1ed779f4c8ee62): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:13 crc kubenswrapper[4968]: E0218 15:33:13.730041 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_openshift-marketplace(62b19fba-8472-4001-b221-2d4f0dd1d36e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_openshift-marketplace(62b19fba-8472-4001-b221-2d4f0dd1d36e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_openshift-marketplace_62b19fba-8472-4001-b221-2d4f0dd1d36e_0(472cb6366899081d28bf55d0c34514db2bfe0409ca2da653ec1ed779f4c8ee62): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" podUID="62b19fba-8472-4001-b221-2d4f0dd1d36e" Feb 18 15:33:14 crc kubenswrapper[4968]: I0218 15:33:14.387863 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2qq8_2d01781a-6a78-49a2-80c7-9ac02c810e3f/kube-multus/2.log" Feb 18 15:33:14 crc kubenswrapper[4968]: I0218 15:33:14.388008 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:14 crc kubenswrapper[4968]: I0218 15:33:14.387994 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2qq8" event={"ID":"2d01781a-6a78-49a2-80c7-9ac02c810e3f","Type":"ContainerStarted","Data":"bc0fe7fb039fe97b8dae57277aefc8d846989086f3c4ea2f04e884183d744eeb"} Feb 18 15:33:14 crc kubenswrapper[4968]: I0218 15:33:14.388648 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:14 crc kubenswrapper[4968]: E0218 15:33:14.434314 4968 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_openshift-marketplace_62b19fba-8472-4001-b221-2d4f0dd1d36e_0(698cd1de3bfa9102f3a801bbfa942f482dda2d3316300ef75bc0f2bf4bd4758a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 18 15:33:14 crc kubenswrapper[4968]: E0218 15:33:14.434889 4968 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_openshift-marketplace_62b19fba-8472-4001-b221-2d4f0dd1d36e_0(698cd1de3bfa9102f3a801bbfa942f482dda2d3316300ef75bc0f2bf4bd4758a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:14 crc kubenswrapper[4968]: E0218 15:33:14.434946 4968 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_openshift-marketplace_62b19fba-8472-4001-b221-2d4f0dd1d36e_0(698cd1de3bfa9102f3a801bbfa942f482dda2d3316300ef75bc0f2bf4bd4758a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:14 crc kubenswrapper[4968]: E0218 15:33:14.435064 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_openshift-marketplace(62b19fba-8472-4001-b221-2d4f0dd1d36e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_openshift-marketplace(62b19fba-8472-4001-b221-2d4f0dd1d36e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_openshift-marketplace_62b19fba-8472-4001-b221-2d4f0dd1d36e_0(698cd1de3bfa9102f3a801bbfa942f482dda2d3316300ef75bc0f2bf4bd4758a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" podUID="62b19fba-8472-4001-b221-2d4f0dd1d36e" Feb 18 15:33:17 crc kubenswrapper[4968]: I0218 15:33:17.296243 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-g7ffz" Feb 18 15:33:29 crc kubenswrapper[4968]: I0218 15:33:29.230588 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:29 crc kubenswrapper[4968]: I0218 15:33:29.231803 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:29 crc kubenswrapper[4968]: I0218 15:33:29.503278 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8"] Feb 18 15:33:30 crc kubenswrapper[4968]: I0218 15:33:30.498389 4968 generic.go:334] "Generic (PLEG): container finished" podID="62b19fba-8472-4001-b221-2d4f0dd1d36e" containerID="d004eafa74d86d70a6af6b826381ce30c368b56f7e9e2c75da2fb5d9618ffe1e" exitCode=0 Feb 18 15:33:30 crc kubenswrapper[4968]: I0218 15:33:30.498452 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" event={"ID":"62b19fba-8472-4001-b221-2d4f0dd1d36e","Type":"ContainerDied","Data":"d004eafa74d86d70a6af6b826381ce30c368b56f7e9e2c75da2fb5d9618ffe1e"} Feb 18 15:33:30 crc kubenswrapper[4968]: I0218 15:33:30.498528 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" event={"ID":"62b19fba-8472-4001-b221-2d4f0dd1d36e","Type":"ContainerStarted","Data":"a7c47d95bfcff244fdf6fc4badc924640e28a66aa717fc55c48f0795146d2501"} Feb 18 15:33:32 crc kubenswrapper[4968]: I0218 15:33:32.521685 4968 generic.go:334] "Generic (PLEG): container finished" podID="62b19fba-8472-4001-b221-2d4f0dd1d36e" containerID="0d5a9acef3d7ffd8d5f4ff2e62c9a2e7618246f5fcd1e3a1541de075711b0136" exitCode=0 Feb 18 15:33:32 crc kubenswrapper[4968]: I0218 15:33:32.521907 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" event={"ID":"62b19fba-8472-4001-b221-2d4f0dd1d36e","Type":"ContainerDied","Data":"0d5a9acef3d7ffd8d5f4ff2e62c9a2e7618246f5fcd1e3a1541de075711b0136"} Feb 18 15:33:33 crc kubenswrapper[4968]: I0218 15:33:33.532694 4968 generic.go:334] "Generic (PLEG): container finished" podID="62b19fba-8472-4001-b221-2d4f0dd1d36e" containerID="d2d0e81a020af970762380a547b00bf01abb93202082823c479832fc349f11e9" exitCode=0 Feb 18 15:33:33 crc kubenswrapper[4968]: I0218 15:33:33.532823 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" event={"ID":"62b19fba-8472-4001-b221-2d4f0dd1d36e","Type":"ContainerDied","Data":"d2d0e81a020af970762380a547b00bf01abb93202082823c479832fc349f11e9"} Feb 18 15:33:34 crc kubenswrapper[4968]: I0218 15:33:34.826308 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:34 crc kubenswrapper[4968]: I0218 15:33:34.962038 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cggc\" (UniqueName: \"kubernetes.io/projected/62b19fba-8472-4001-b221-2d4f0dd1d36e-kube-api-access-4cggc\") pod \"62b19fba-8472-4001-b221-2d4f0dd1d36e\" (UID: \"62b19fba-8472-4001-b221-2d4f0dd1d36e\") " Feb 18 15:33:34 crc kubenswrapper[4968]: I0218 15:33:34.962129 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/62b19fba-8472-4001-b221-2d4f0dd1d36e-util\") pod \"62b19fba-8472-4001-b221-2d4f0dd1d36e\" (UID: \"62b19fba-8472-4001-b221-2d4f0dd1d36e\") " Feb 18 15:33:34 crc kubenswrapper[4968]: I0218 15:33:34.962195 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/62b19fba-8472-4001-b221-2d4f0dd1d36e-bundle\") pod \"62b19fba-8472-4001-b221-2d4f0dd1d36e\" (UID: \"62b19fba-8472-4001-b221-2d4f0dd1d36e\") " Feb 18 15:33:34 crc kubenswrapper[4968]: I0218 15:33:34.965178 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62b19fba-8472-4001-b221-2d4f0dd1d36e-bundle" (OuterVolumeSpecName: "bundle") pod "62b19fba-8472-4001-b221-2d4f0dd1d36e" (UID: "62b19fba-8472-4001-b221-2d4f0dd1d36e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:33:34 crc kubenswrapper[4968]: I0218 15:33:34.970926 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62b19fba-8472-4001-b221-2d4f0dd1d36e-kube-api-access-4cggc" (OuterVolumeSpecName: "kube-api-access-4cggc") pod "62b19fba-8472-4001-b221-2d4f0dd1d36e" (UID: "62b19fba-8472-4001-b221-2d4f0dd1d36e"). InnerVolumeSpecName "kube-api-access-4cggc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:33:34 crc kubenswrapper[4968]: I0218 15:33:34.993387 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62b19fba-8472-4001-b221-2d4f0dd1d36e-util" (OuterVolumeSpecName: "util") pod "62b19fba-8472-4001-b221-2d4f0dd1d36e" (UID: "62b19fba-8472-4001-b221-2d4f0dd1d36e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:33:35 crc kubenswrapper[4968]: I0218 15:33:35.063590 4968 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/62b19fba-8472-4001-b221-2d4f0dd1d36e-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:33:35 crc kubenswrapper[4968]: I0218 15:33:35.063630 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cggc\" (UniqueName: \"kubernetes.io/projected/62b19fba-8472-4001-b221-2d4f0dd1d36e-kube-api-access-4cggc\") on node \"crc\" DevicePath \"\"" Feb 18 15:33:35 crc kubenswrapper[4968]: I0218 15:33:35.063644 4968 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/62b19fba-8472-4001-b221-2d4f0dd1d36e-util\") on node \"crc\" DevicePath \"\"" Feb 18 15:33:35 crc kubenswrapper[4968]: I0218 15:33:35.551482 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" event={"ID":"62b19fba-8472-4001-b221-2d4f0dd1d36e","Type":"ContainerDied","Data":"a7c47d95bfcff244fdf6fc4badc924640e28a66aa717fc55c48f0795146d2501"} Feb 18 15:33:35 crc kubenswrapper[4968]: I0218 15:33:35.551542 4968 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7c47d95bfcff244fdf6fc4badc924640e28a66aa717fc55c48f0795146d2501" Feb 18 15:33:35 crc kubenswrapper[4968]: I0218 15:33:35.551599 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8" Feb 18 15:33:44 crc kubenswrapper[4968]: I0218 15:33:44.367698 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:33:44 crc kubenswrapper[4968]: I0218 15:33:44.368176 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.593054 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-d8lx8"] Feb 18 15:33:46 crc kubenswrapper[4968]: E0218 15:33:46.593344 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b19fba-8472-4001-b221-2d4f0dd1d36e" containerName="extract" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.593362 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b19fba-8472-4001-b221-2d4f0dd1d36e" containerName="extract" Feb 18 15:33:46 crc kubenswrapper[4968]: E0218 15:33:46.593379 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b19fba-8472-4001-b221-2d4f0dd1d36e" containerName="util" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.593389 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b19fba-8472-4001-b221-2d4f0dd1d36e" containerName="util" Feb 18 15:33:46 crc kubenswrapper[4968]: E0218 15:33:46.593411 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b19fba-8472-4001-b221-2d4f0dd1d36e" containerName="pull" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.593422 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b19fba-8472-4001-b221-2d4f0dd1d36e" containerName="pull" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.593545 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b19fba-8472-4001-b221-2d4f0dd1d36e" containerName="extract" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.594021 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-d8lx8" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.597149 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.597269 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.597476 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-v2mmv" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.603744 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-d8lx8"] Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.627253 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42n9v\" (UniqueName: \"kubernetes.io/projected/96458238-fdec-497e-988d-0d6d6a59af36-kube-api-access-42n9v\") pod \"obo-prometheus-operator-68bc856cb9-d8lx8\" (UID: \"96458238-fdec-497e-988d-0d6d6a59af36\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-d8lx8" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.705462 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw"] Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.706195 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.708129 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.708522 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-d6l4t" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.717598 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x"] Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.718485 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.728163 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/de5c536d-8040-4425-8bf1-5deda35fe0c6-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw\" (UID: \"de5c536d-8040-4425-8bf1-5deda35fe0c6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.728216 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/de5c536d-8040-4425-8bf1-5deda35fe0c6-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw\" (UID: \"de5c536d-8040-4425-8bf1-5deda35fe0c6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.728271 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42n9v\" (UniqueName: \"kubernetes.io/projected/96458238-fdec-497e-988d-0d6d6a59af36-kube-api-access-42n9v\") pod \"obo-prometheus-operator-68bc856cb9-d8lx8\" (UID: \"96458238-fdec-497e-988d-0d6d6a59af36\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-d8lx8" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.728877 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw"] Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.733170 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x"] Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.765774 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42n9v\" (UniqueName: \"kubernetes.io/projected/96458238-fdec-497e-988d-0d6d6a59af36-kube-api-access-42n9v\") pod \"obo-prometheus-operator-68bc856cb9-d8lx8\" (UID: \"96458238-fdec-497e-988d-0d6d6a59af36\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-d8lx8" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.829103 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/de5c536d-8040-4425-8bf1-5deda35fe0c6-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw\" (UID: \"de5c536d-8040-4425-8bf1-5deda35fe0c6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.829462 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d21d7e50-1c72-4fc6-ad51-3603cccbce41-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-57669fbcb7-8777x\" (UID: \"d21d7e50-1c72-4fc6-ad51-3603cccbce41\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.829502 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d21d7e50-1c72-4fc6-ad51-3603cccbce41-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-57669fbcb7-8777x\" (UID: \"d21d7e50-1c72-4fc6-ad51-3603cccbce41\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.829531 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/de5c536d-8040-4425-8bf1-5deda35fe0c6-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw\" (UID: \"de5c536d-8040-4425-8bf1-5deda35fe0c6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.844281 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/de5c536d-8040-4425-8bf1-5deda35fe0c6-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw\" (UID: \"de5c536d-8040-4425-8bf1-5deda35fe0c6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.847953 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/de5c536d-8040-4425-8bf1-5deda35fe0c6-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw\" (UID: \"de5c536d-8040-4425-8bf1-5deda35fe0c6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.859107 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-mg5x7"] Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.859896 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.866000 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-2m59d" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.867023 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.876039 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-mg5x7"] Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.916336 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-d8lx8" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.930501 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/c1f3704e-f185-49f7-8d17-226dd2f42061-observability-operator-tls\") pod \"observability-operator-59bdc8b94-mg5x7\" (UID: \"c1f3704e-f185-49f7-8d17-226dd2f42061\") " pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.930568 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfs26\" (UniqueName: \"kubernetes.io/projected/c1f3704e-f185-49f7-8d17-226dd2f42061-kube-api-access-mfs26\") pod \"observability-operator-59bdc8b94-mg5x7\" (UID: \"c1f3704e-f185-49f7-8d17-226dd2f42061\") " pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.930610 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d21d7e50-1c72-4fc6-ad51-3603cccbce41-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-57669fbcb7-8777x\" (UID: \"d21d7e50-1c72-4fc6-ad51-3603cccbce41\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.930639 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d21d7e50-1c72-4fc6-ad51-3603cccbce41-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-57669fbcb7-8777x\" (UID: \"d21d7e50-1c72-4fc6-ad51-3603cccbce41\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.933439 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d21d7e50-1c72-4fc6-ad51-3603cccbce41-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-57669fbcb7-8777x\" (UID: \"d21d7e50-1c72-4fc6-ad51-3603cccbce41\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x" Feb 18 15:33:46 crc kubenswrapper[4968]: I0218 15:33:46.933726 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d21d7e50-1c72-4fc6-ad51-3603cccbce41-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-57669fbcb7-8777x\" (UID: \"d21d7e50-1c72-4fc6-ad51-3603cccbce41\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.031753 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-7f95q"] Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.032458 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-7f95q" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.033073 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.033299 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfs26\" (UniqueName: \"kubernetes.io/projected/c1f3704e-f185-49f7-8d17-226dd2f42061-kube-api-access-mfs26\") pod \"observability-operator-59bdc8b94-mg5x7\" (UID: \"c1f3704e-f185-49f7-8d17-226dd2f42061\") " pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.033377 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/c1f3704e-f185-49f7-8d17-226dd2f42061-observability-operator-tls\") pod \"observability-operator-59bdc8b94-mg5x7\" (UID: \"c1f3704e-f185-49f7-8d17-226dd2f42061\") " pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.043873 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/c1f3704e-f185-49f7-8d17-226dd2f42061-observability-operator-tls\") pod \"observability-operator-59bdc8b94-mg5x7\" (UID: \"c1f3704e-f185-49f7-8d17-226dd2f42061\") " pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.045684 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-x5fpj" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.046138 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.052439 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-7f95q"] Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.074480 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfs26\" (UniqueName: \"kubernetes.io/projected/c1f3704e-f185-49f7-8d17-226dd2f42061-kube-api-access-mfs26\") pod \"observability-operator-59bdc8b94-mg5x7\" (UID: \"c1f3704e-f185-49f7-8d17-226dd2f42061\") " pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.134258 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/ff69eec4-70ef-4b36-ab53-93131c8d15ce-openshift-service-ca\") pod \"perses-operator-5bf474d74f-7f95q\" (UID: \"ff69eec4-70ef-4b36-ab53-93131c8d15ce\") " pod="openshift-operators/perses-operator-5bf474d74f-7f95q" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.134899 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6v6f\" (UniqueName: \"kubernetes.io/projected/ff69eec4-70ef-4b36-ab53-93131c8d15ce-kube-api-access-d6v6f\") pod \"perses-operator-5bf474d74f-7f95q\" (UID: \"ff69eec4-70ef-4b36-ab53-93131c8d15ce\") " pod="openshift-operators/perses-operator-5bf474d74f-7f95q" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.180453 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-d8lx8"] Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.186404 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.238204 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6v6f\" (UniqueName: \"kubernetes.io/projected/ff69eec4-70ef-4b36-ab53-93131c8d15ce-kube-api-access-d6v6f\") pod \"perses-operator-5bf474d74f-7f95q\" (UID: \"ff69eec4-70ef-4b36-ab53-93131c8d15ce\") " pod="openshift-operators/perses-operator-5bf474d74f-7f95q" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.238268 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/ff69eec4-70ef-4b36-ab53-93131c8d15ce-openshift-service-ca\") pod \"perses-operator-5bf474d74f-7f95q\" (UID: \"ff69eec4-70ef-4b36-ab53-93131c8d15ce\") " pod="openshift-operators/perses-operator-5bf474d74f-7f95q" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.242101 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/ff69eec4-70ef-4b36-ab53-93131c8d15ce-openshift-service-ca\") pod \"perses-operator-5bf474d74f-7f95q\" (UID: \"ff69eec4-70ef-4b36-ab53-93131c8d15ce\") " pod="openshift-operators/perses-operator-5bf474d74f-7f95q" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.271816 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6v6f\" (UniqueName: \"kubernetes.io/projected/ff69eec4-70ef-4b36-ab53-93131c8d15ce-kube-api-access-d6v6f\") pod \"perses-operator-5bf474d74f-7f95q\" (UID: \"ff69eec4-70ef-4b36-ab53-93131c8d15ce\") " pod="openshift-operators/perses-operator-5bf474d74f-7f95q" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.275193 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x"] Feb 18 15:33:47 crc kubenswrapper[4968]: W0218 15:33:47.298301 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd21d7e50_1c72_4fc6_ad51_3603cccbce41.slice/crio-16a8d1658c7b74d7938dc378fa754b3e525dd534982d7a83a3ce4e4aaee7ccbf WatchSource:0}: Error finding container 16a8d1658c7b74d7938dc378fa754b3e525dd534982d7a83a3ce4e4aaee7ccbf: Status 404 returned error can't find the container with id 16a8d1658c7b74d7938dc378fa754b3e525dd534982d7a83a3ce4e4aaee7ccbf Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.356974 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw"] Feb 18 15:33:47 crc kubenswrapper[4968]: W0218 15:33:47.365591 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde5c536d_8040_4425_8bf1_5deda35fe0c6.slice/crio-e938af1f9d383481d64ab9e3a21a01c20e7b35e9877f730cd06f7c2942c3c036 WatchSource:0}: Error finding container e938af1f9d383481d64ab9e3a21a01c20e7b35e9877f730cd06f7c2942c3c036: Status 404 returned error can't find the container with id e938af1f9d383481d64ab9e3a21a01c20e7b35e9877f730cd06f7c2942c3c036 Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.379771 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-7f95q" Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.466380 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-mg5x7"] Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.576958 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-7f95q"] Feb 18 15:33:47 crc kubenswrapper[4968]: W0218 15:33:47.583731 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff69eec4_70ef_4b36_ab53_93131c8d15ce.slice/crio-25d396b3a4deb70d5733aa555ba2cb7406e496fcdb080aebd063ddf10bf9fcb2 WatchSource:0}: Error finding container 25d396b3a4deb70d5733aa555ba2cb7406e496fcdb080aebd063ddf10bf9fcb2: Status 404 returned error can't find the container with id 25d396b3a4deb70d5733aa555ba2cb7406e496fcdb080aebd063ddf10bf9fcb2 Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.620115 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" event={"ID":"c1f3704e-f185-49f7-8d17-226dd2f42061","Type":"ContainerStarted","Data":"c1cca06359ca50c7646f9cc1fddc46889d92caf0587907941f1d1fcefa2fdf7f"} Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.621313 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x" event={"ID":"d21d7e50-1c72-4fc6-ad51-3603cccbce41","Type":"ContainerStarted","Data":"16a8d1658c7b74d7938dc378fa754b3e525dd534982d7a83a3ce4e4aaee7ccbf"} Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.622110 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-d8lx8" event={"ID":"96458238-fdec-497e-988d-0d6d6a59af36","Type":"ContainerStarted","Data":"a89e1307730bd3f693c07f70fa787f6ee55a44bf95cf1b1a306813382c310db6"} Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.623120 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-7f95q" event={"ID":"ff69eec4-70ef-4b36-ab53-93131c8d15ce","Type":"ContainerStarted","Data":"25d396b3a4deb70d5733aa555ba2cb7406e496fcdb080aebd063ddf10bf9fcb2"} Feb 18 15:33:47 crc kubenswrapper[4968]: I0218 15:33:47.623969 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw" event={"ID":"de5c536d-8040-4425-8bf1-5deda35fe0c6","Type":"ContainerStarted","Data":"e938af1f9d383481d64ab9e3a21a01c20e7b35e9877f730cd06f7c2942c3c036"} Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.736860 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-d8lx8" event={"ID":"96458238-fdec-497e-988d-0d6d6a59af36","Type":"ContainerStarted","Data":"19d98faf806faa34bcf57bd007fa641ad4e2542e9f64665dc78338ce501880a9"} Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.739359 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-7f95q" event={"ID":"ff69eec4-70ef-4b36-ab53-93131c8d15ce","Type":"ContainerStarted","Data":"a1c0f9df43ae269b2763fd1fcebb0e292c139f4a2450024b29d33a3d984184b4"} Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.739484 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-7f95q" Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.741869 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw" event={"ID":"de5c536d-8040-4425-8bf1-5deda35fe0c6","Type":"ContainerStarted","Data":"7da6ccd91d9c31f512073b1edb64bfd6a0bdba6e6024b75280a1988a74340b36"} Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.743723 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" event={"ID":"c1f3704e-f185-49f7-8d17-226dd2f42061","Type":"ContainerStarted","Data":"fe74106be7b4ecd94e4a8f65c320995bf29d803151bf8e527ef47c12c9982212"} Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.743907 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.747084 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x" event={"ID":"d21d7e50-1c72-4fc6-ad51-3603cccbce41","Type":"ContainerStarted","Data":"667a7168126e48a640e60160dad335bd09676dedb84b1fa95f8499e3483d2653"} Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.755294 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-d8lx8" podStartSLOduration=2.403899739 podStartE2EDuration="14.755176019s" podCreationTimestamp="2026-02-18 15:33:46 +0000 UTC" firstStartedPulling="2026-02-18 15:33:47.209446043 +0000 UTC m=+726.594890905" lastFinishedPulling="2026-02-18 15:33:59.560722313 +0000 UTC m=+738.946167185" observedRunningTime="2026-02-18 15:34:00.751495496 +0000 UTC m=+740.136940358" watchObservedRunningTime="2026-02-18 15:34:00.755176019 +0000 UTC m=+740.140620891" Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.762576 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.769559 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-7f95q" podStartSLOduration=1.825315009 podStartE2EDuration="13.769541991s" podCreationTimestamp="2026-02-18 15:33:47 +0000 UTC" firstStartedPulling="2026-02-18 15:33:47.585901545 +0000 UTC m=+726.971346397" lastFinishedPulling="2026-02-18 15:33:59.530128507 +0000 UTC m=+738.915573379" observedRunningTime="2026-02-18 15:34:00.767924376 +0000 UTC m=+740.153369238" watchObservedRunningTime="2026-02-18 15:34:00.769541991 +0000 UTC m=+740.154986863" Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.802507 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-mg5x7" podStartSLOduration=2.714100927 podStartE2EDuration="14.802489073s" podCreationTimestamp="2026-02-18 15:33:46 +0000 UTC" firstStartedPulling="2026-02-18 15:33:47.471223176 +0000 UTC m=+726.856668038" lastFinishedPulling="2026-02-18 15:33:59.559611192 +0000 UTC m=+738.945056184" observedRunningTime="2026-02-18 15:34:00.799860609 +0000 UTC m=+740.185305491" watchObservedRunningTime="2026-02-18 15:34:00.802489073 +0000 UTC m=+740.187933935" Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.823069 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-8777x" podStartSLOduration=2.6123209899999997 podStartE2EDuration="14.823028357s" podCreationTimestamp="2026-02-18 15:33:46 +0000 UTC" firstStartedPulling="2026-02-18 15:33:47.320933262 +0000 UTC m=+726.706378124" lastFinishedPulling="2026-02-18 15:33:59.531640629 +0000 UTC m=+738.917085491" observedRunningTime="2026-02-18 15:34:00.820442805 +0000 UTC m=+740.205887667" watchObservedRunningTime="2026-02-18 15:34:00.823028357 +0000 UTC m=+740.208473229" Feb 18 15:34:00 crc kubenswrapper[4968]: I0218 15:34:00.877689 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw" podStartSLOduration=2.734747245 podStartE2EDuration="14.877668606s" podCreationTimestamp="2026-02-18 15:33:46 +0000 UTC" firstStartedPulling="2026-02-18 15:33:47.370582131 +0000 UTC m=+726.756026993" lastFinishedPulling="2026-02-18 15:33:59.513503492 +0000 UTC m=+738.898948354" observedRunningTime="2026-02-18 15:34:00.869156738 +0000 UTC m=+740.254601610" watchObservedRunningTime="2026-02-18 15:34:00.877668606 +0000 UTC m=+740.263113488" Feb 18 15:34:07 crc kubenswrapper[4968]: I0218 15:34:07.382889 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-7f95q" Feb 18 15:34:14 crc kubenswrapper[4968]: I0218 15:34:14.367622 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:34:14 crc kubenswrapper[4968]: I0218 15:34:14.368309 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.244388 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k"] Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.245954 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.259804 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.264030 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k"] Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.292018 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xczsd\" (UniqueName: \"kubernetes.io/projected/67350a11-27a5-46e6-bab7-32c8a7bd74df-kube-api-access-xczsd\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k\" (UID: \"67350a11-27a5-46e6-bab7-32c8a7bd74df\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.292092 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/67350a11-27a5-46e6-bab7-32c8a7bd74df-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k\" (UID: \"67350a11-27a5-46e6-bab7-32c8a7bd74df\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.292121 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/67350a11-27a5-46e6-bab7-32c8a7bd74df-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k\" (UID: \"67350a11-27a5-46e6-bab7-32c8a7bd74df\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.393465 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xczsd\" (UniqueName: \"kubernetes.io/projected/67350a11-27a5-46e6-bab7-32c8a7bd74df-kube-api-access-xczsd\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k\" (UID: \"67350a11-27a5-46e6-bab7-32c8a7bd74df\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.393542 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/67350a11-27a5-46e6-bab7-32c8a7bd74df-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k\" (UID: \"67350a11-27a5-46e6-bab7-32c8a7bd74df\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.393567 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/67350a11-27a5-46e6-bab7-32c8a7bd74df-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k\" (UID: \"67350a11-27a5-46e6-bab7-32c8a7bd74df\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.394258 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/67350a11-27a5-46e6-bab7-32c8a7bd74df-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k\" (UID: \"67350a11-27a5-46e6-bab7-32c8a7bd74df\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.394383 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/67350a11-27a5-46e6-bab7-32c8a7bd74df-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k\" (UID: \"67350a11-27a5-46e6-bab7-32c8a7bd74df\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.416663 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xczsd\" (UniqueName: \"kubernetes.io/projected/67350a11-27a5-46e6-bab7-32c8a7bd74df-kube-api-access-xczsd\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k\" (UID: \"67350a11-27a5-46e6-bab7-32c8a7bd74df\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.565232 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.777102 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k"] Feb 18 15:34:23 crc kubenswrapper[4968]: I0218 15:34:23.888552 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" event={"ID":"67350a11-27a5-46e6-bab7-32c8a7bd74df","Type":"ContainerStarted","Data":"8dfe4dfe6bd81c0ae319503ee4dabe457b228add4617a98fd3c8f6ab3f4c8900"} Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.002395 4968 generic.go:334] "Generic (PLEG): container finished" podID="67350a11-27a5-46e6-bab7-32c8a7bd74df" containerID="312657328c9a00d40c774266a552c1cb3675b4167125cc87f5f92c4fe3810793" exitCode=0 Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.002437 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" event={"ID":"67350a11-27a5-46e6-bab7-32c8a7bd74df","Type":"ContainerDied","Data":"312657328c9a00d40c774266a552c1cb3675b4167125cc87f5f92c4fe3810793"} Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.606185 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cwp6r"] Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.607461 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.623935 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cwp6r"] Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.691044 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-catalog-content\") pod \"redhat-operators-cwp6r\" (UID: \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\") " pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.691099 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwj59\" (UniqueName: \"kubernetes.io/projected/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-kube-api-access-kwj59\") pod \"redhat-operators-cwp6r\" (UID: \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\") " pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.691131 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-utilities\") pod \"redhat-operators-cwp6r\" (UID: \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\") " pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.791961 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-utilities\") pod \"redhat-operators-cwp6r\" (UID: \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\") " pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.792377 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-catalog-content\") pod \"redhat-operators-cwp6r\" (UID: \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\") " pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.792569 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-utilities\") pod \"redhat-operators-cwp6r\" (UID: \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\") " pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.792578 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwj59\" (UniqueName: \"kubernetes.io/projected/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-kube-api-access-kwj59\") pod \"redhat-operators-cwp6r\" (UID: \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\") " pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.792740 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-catalog-content\") pod \"redhat-operators-cwp6r\" (UID: \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\") " pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.816200 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwj59\" (UniqueName: \"kubernetes.io/projected/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-kube-api-access-kwj59\") pod \"redhat-operators-cwp6r\" (UID: \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\") " pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:25 crc kubenswrapper[4968]: I0218 15:34:25.928614 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:26 crc kubenswrapper[4968]: I0218 15:34:26.224935 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cwp6r"] Feb 18 15:34:26 crc kubenswrapper[4968]: I0218 15:34:26.745256 4968 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 18 15:34:27 crc kubenswrapper[4968]: I0218 15:34:27.026159 4968 generic.go:334] "Generic (PLEG): container finished" podID="67350a11-27a5-46e6-bab7-32c8a7bd74df" containerID="563b4952fff51c35075f424df9c19fb209a1cb97957ef27deaae5b19129c00a2" exitCode=0 Feb 18 15:34:27 crc kubenswrapper[4968]: I0218 15:34:27.026194 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" event={"ID":"67350a11-27a5-46e6-bab7-32c8a7bd74df","Type":"ContainerDied","Data":"563b4952fff51c35075f424df9c19fb209a1cb97957ef27deaae5b19129c00a2"} Feb 18 15:34:27 crc kubenswrapper[4968]: I0218 15:34:27.027955 4968 generic.go:334] "Generic (PLEG): container finished" podID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" containerID="0390185bdf60d40a58dee59b18d551aa98b8f34a590a8a22f95c7bd00666e515" exitCode=0 Feb 18 15:34:27 crc kubenswrapper[4968]: I0218 15:34:27.027987 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwp6r" event={"ID":"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642","Type":"ContainerDied","Data":"0390185bdf60d40a58dee59b18d551aa98b8f34a590a8a22f95c7bd00666e515"} Feb 18 15:34:27 crc kubenswrapper[4968]: I0218 15:34:27.028017 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwp6r" event={"ID":"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642","Type":"ContainerStarted","Data":"934aaa129af7b3c28fd5ba07a21e2380e7cb7e56bc808db529125a8d07de7111"} Feb 18 15:34:28 crc kubenswrapper[4968]: I0218 15:34:28.035654 4968 generic.go:334] "Generic (PLEG): container finished" podID="67350a11-27a5-46e6-bab7-32c8a7bd74df" containerID="d2fecb979f0f2f20db040b29e1359f59c5afb7c994edb57c0b87b317d4503842" exitCode=0 Feb 18 15:34:28 crc kubenswrapper[4968]: I0218 15:34:28.035737 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" event={"ID":"67350a11-27a5-46e6-bab7-32c8a7bd74df","Type":"ContainerDied","Data":"d2fecb979f0f2f20db040b29e1359f59c5afb7c994edb57c0b87b317d4503842"} Feb 18 15:34:28 crc kubenswrapper[4968]: I0218 15:34:28.038534 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwp6r" event={"ID":"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642","Type":"ContainerStarted","Data":"8e1baebd50335277fde9216691678ac2d8e03a2f5eecbb8cb58de82bafe4a767"} Feb 18 15:34:28 crc kubenswrapper[4968]: I0218 15:34:28.812214 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c82lb"] Feb 18 15:34:28 crc kubenswrapper[4968]: I0218 15:34:28.814292 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:28 crc kubenswrapper[4968]: I0218 15:34:28.829769 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c82lb"] Feb 18 15:34:28 crc kubenswrapper[4968]: I0218 15:34:28.930994 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnbwc\" (UniqueName: \"kubernetes.io/projected/d05ee3b8-c378-467d-a0cd-d58121ae17a9-kube-api-access-fnbwc\") pod \"redhat-marketplace-c82lb\" (UID: \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\") " pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:28 crc kubenswrapper[4968]: I0218 15:34:28.931060 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05ee3b8-c378-467d-a0cd-d58121ae17a9-catalog-content\") pod \"redhat-marketplace-c82lb\" (UID: \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\") " pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:28 crc kubenswrapper[4968]: I0218 15:34:28.931198 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05ee3b8-c378-467d-a0cd-d58121ae17a9-utilities\") pod \"redhat-marketplace-c82lb\" (UID: \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\") " pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.032992 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05ee3b8-c378-467d-a0cd-d58121ae17a9-catalog-content\") pod \"redhat-marketplace-c82lb\" (UID: \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\") " pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.033054 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnbwc\" (UniqueName: \"kubernetes.io/projected/d05ee3b8-c378-467d-a0cd-d58121ae17a9-kube-api-access-fnbwc\") pod \"redhat-marketplace-c82lb\" (UID: \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\") " pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.033152 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05ee3b8-c378-467d-a0cd-d58121ae17a9-utilities\") pod \"redhat-marketplace-c82lb\" (UID: \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\") " pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.033536 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05ee3b8-c378-467d-a0cd-d58121ae17a9-catalog-content\") pod \"redhat-marketplace-c82lb\" (UID: \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\") " pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.033691 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05ee3b8-c378-467d-a0cd-d58121ae17a9-utilities\") pod \"redhat-marketplace-c82lb\" (UID: \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\") " pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.044604 4968 generic.go:334] "Generic (PLEG): container finished" podID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" containerID="8e1baebd50335277fde9216691678ac2d8e03a2f5eecbb8cb58de82bafe4a767" exitCode=0 Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.045429 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwp6r" event={"ID":"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642","Type":"ContainerDied","Data":"8e1baebd50335277fde9216691678ac2d8e03a2f5eecbb8cb58de82bafe4a767"} Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.073237 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnbwc\" (UniqueName: \"kubernetes.io/projected/d05ee3b8-c378-467d-a0cd-d58121ae17a9-kube-api-access-fnbwc\") pod \"redhat-marketplace-c82lb\" (UID: \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\") " pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.143925 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.378540 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.409795 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c82lb"] Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.540693 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xczsd\" (UniqueName: \"kubernetes.io/projected/67350a11-27a5-46e6-bab7-32c8a7bd74df-kube-api-access-xczsd\") pod \"67350a11-27a5-46e6-bab7-32c8a7bd74df\" (UID: \"67350a11-27a5-46e6-bab7-32c8a7bd74df\") " Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.540769 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/67350a11-27a5-46e6-bab7-32c8a7bd74df-util\") pod \"67350a11-27a5-46e6-bab7-32c8a7bd74df\" (UID: \"67350a11-27a5-46e6-bab7-32c8a7bd74df\") " Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.540822 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/67350a11-27a5-46e6-bab7-32c8a7bd74df-bundle\") pod \"67350a11-27a5-46e6-bab7-32c8a7bd74df\" (UID: \"67350a11-27a5-46e6-bab7-32c8a7bd74df\") " Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.541391 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67350a11-27a5-46e6-bab7-32c8a7bd74df-bundle" (OuterVolumeSpecName: "bundle") pod "67350a11-27a5-46e6-bab7-32c8a7bd74df" (UID: "67350a11-27a5-46e6-bab7-32c8a7bd74df"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.544345 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67350a11-27a5-46e6-bab7-32c8a7bd74df-kube-api-access-xczsd" (OuterVolumeSpecName: "kube-api-access-xczsd") pod "67350a11-27a5-46e6-bab7-32c8a7bd74df" (UID: "67350a11-27a5-46e6-bab7-32c8a7bd74df"). InnerVolumeSpecName "kube-api-access-xczsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.641820 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xczsd\" (UniqueName: \"kubernetes.io/projected/67350a11-27a5-46e6-bab7-32c8a7bd74df-kube-api-access-xczsd\") on node \"crc\" DevicePath \"\"" Feb 18 15:34:29 crc kubenswrapper[4968]: I0218 15:34:29.641853 4968 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/67350a11-27a5-46e6-bab7-32c8a7bd74df-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:34:30 crc kubenswrapper[4968]: I0218 15:34:30.052572 4968 generic.go:334] "Generic (PLEG): container finished" podID="d05ee3b8-c378-467d-a0cd-d58121ae17a9" containerID="3c88df8ac907236c1332e38a2c8cd86f0a5d05dedc0b91bea650e06e51e78ecc" exitCode=0 Feb 18 15:34:30 crc kubenswrapper[4968]: I0218 15:34:30.052635 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c82lb" event={"ID":"d05ee3b8-c378-467d-a0cd-d58121ae17a9","Type":"ContainerDied","Data":"3c88df8ac907236c1332e38a2c8cd86f0a5d05dedc0b91bea650e06e51e78ecc"} Feb 18 15:34:30 crc kubenswrapper[4968]: I0218 15:34:30.053055 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c82lb" event={"ID":"d05ee3b8-c378-467d-a0cd-d58121ae17a9","Type":"ContainerStarted","Data":"0c39aa692a2a7e6ba8530203a715607b50232888a22b231d78fc15cf53843f7d"} Feb 18 15:34:30 crc kubenswrapper[4968]: I0218 15:34:30.055403 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" event={"ID":"67350a11-27a5-46e6-bab7-32c8a7bd74df","Type":"ContainerDied","Data":"8dfe4dfe6bd81c0ae319503ee4dabe457b228add4617a98fd3c8f6ab3f4c8900"} Feb 18 15:34:30 crc kubenswrapper[4968]: I0218 15:34:30.055438 4968 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dfe4dfe6bd81c0ae319503ee4dabe457b228add4617a98fd3c8f6ab3f4c8900" Feb 18 15:34:30 crc kubenswrapper[4968]: I0218 15:34:30.055500 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k" Feb 18 15:34:30 crc kubenswrapper[4968]: I0218 15:34:30.617877 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67350a11-27a5-46e6-bab7-32c8a7bd74df-util" (OuterVolumeSpecName: "util") pod "67350a11-27a5-46e6-bab7-32c8a7bd74df" (UID: "67350a11-27a5-46e6-bab7-32c8a7bd74df"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:34:30 crc kubenswrapper[4968]: I0218 15:34:30.656633 4968 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/67350a11-27a5-46e6-bab7-32c8a7bd74df-util\") on node \"crc\" DevicePath \"\"" Feb 18 15:34:31 crc kubenswrapper[4968]: I0218 15:34:31.062384 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c82lb" event={"ID":"d05ee3b8-c378-467d-a0cd-d58121ae17a9","Type":"ContainerStarted","Data":"dc2d73943abae31e1b7072dcde3a97504a66056363d924193efd0b4ce8e3fe81"} Feb 18 15:34:31 crc kubenswrapper[4968]: I0218 15:34:31.064313 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwp6r" event={"ID":"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642","Type":"ContainerStarted","Data":"4d984d9c4c914066fcb166329e1de8982a2ffb666b7f8d69c336aecc6423b52b"} Feb 18 15:34:31 crc kubenswrapper[4968]: I0218 15:34:31.105402 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cwp6r" podStartSLOduration=2.517142861 podStartE2EDuration="6.105382065s" podCreationTimestamp="2026-02-18 15:34:25 +0000 UTC" firstStartedPulling="2026-02-18 15:34:27.029157729 +0000 UTC m=+766.414602591" lastFinishedPulling="2026-02-18 15:34:30.617396933 +0000 UTC m=+770.002841795" observedRunningTime="2026-02-18 15:34:31.10125009 +0000 UTC m=+770.486694952" watchObservedRunningTime="2026-02-18 15:34:31.105382065 +0000 UTC m=+770.490826927" Feb 18 15:34:32 crc kubenswrapper[4968]: I0218 15:34:32.072311 4968 generic.go:334] "Generic (PLEG): container finished" podID="d05ee3b8-c378-467d-a0cd-d58121ae17a9" containerID="dc2d73943abae31e1b7072dcde3a97504a66056363d924193efd0b4ce8e3fe81" exitCode=0 Feb 18 15:34:32 crc kubenswrapper[4968]: I0218 15:34:32.072355 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c82lb" event={"ID":"d05ee3b8-c378-467d-a0cd-d58121ae17a9","Type":"ContainerDied","Data":"dc2d73943abae31e1b7072dcde3a97504a66056363d924193efd0b4ce8e3fe81"} Feb 18 15:34:33 crc kubenswrapper[4968]: I0218 15:34:33.079827 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c82lb" event={"ID":"d05ee3b8-c378-467d-a0cd-d58121ae17a9","Type":"ContainerStarted","Data":"b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80"} Feb 18 15:34:33 crc kubenswrapper[4968]: I0218 15:34:33.098270 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c82lb" podStartSLOduration=2.589479452 podStartE2EDuration="5.098210717s" podCreationTimestamp="2026-02-18 15:34:28 +0000 UTC" firstStartedPulling="2026-02-18 15:34:30.054778753 +0000 UTC m=+769.440223625" lastFinishedPulling="2026-02-18 15:34:32.563510028 +0000 UTC m=+771.948954890" observedRunningTime="2026-02-18 15:34:33.096120939 +0000 UTC m=+772.481565801" watchObservedRunningTime="2026-02-18 15:34:33.098210717 +0000 UTC m=+772.483655579" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.233725 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-jlkfr"] Feb 18 15:34:34 crc kubenswrapper[4968]: E0218 15:34:34.234306 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67350a11-27a5-46e6-bab7-32c8a7bd74df" containerName="util" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.234326 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="67350a11-27a5-46e6-bab7-32c8a7bd74df" containerName="util" Feb 18 15:34:34 crc kubenswrapper[4968]: E0218 15:34:34.234349 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67350a11-27a5-46e6-bab7-32c8a7bd74df" containerName="pull" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.234359 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="67350a11-27a5-46e6-bab7-32c8a7bd74df" containerName="pull" Feb 18 15:34:34 crc kubenswrapper[4968]: E0218 15:34:34.234384 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67350a11-27a5-46e6-bab7-32c8a7bd74df" containerName="extract" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.234396 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="67350a11-27a5-46e6-bab7-32c8a7bd74df" containerName="extract" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.234538 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="67350a11-27a5-46e6-bab7-32c8a7bd74df" containerName="extract" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.235048 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-jlkfr" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.236979 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-twtdp" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.236980 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.240364 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.254352 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-jlkfr"] Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.401872 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwc7v\" (UniqueName: \"kubernetes.io/projected/63e09cfe-38d6-44fd-859c-3bc876d8808f-kube-api-access-gwc7v\") pod \"nmstate-operator-694c9596b7-jlkfr\" (UID: \"63e09cfe-38d6-44fd-859c-3bc876d8808f\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-jlkfr" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.503413 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwc7v\" (UniqueName: \"kubernetes.io/projected/63e09cfe-38d6-44fd-859c-3bc876d8808f-kube-api-access-gwc7v\") pod \"nmstate-operator-694c9596b7-jlkfr\" (UID: \"63e09cfe-38d6-44fd-859c-3bc876d8808f\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-jlkfr" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.520591 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwc7v\" (UniqueName: \"kubernetes.io/projected/63e09cfe-38d6-44fd-859c-3bc876d8808f-kube-api-access-gwc7v\") pod \"nmstate-operator-694c9596b7-jlkfr\" (UID: \"63e09cfe-38d6-44fd-859c-3bc876d8808f\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-jlkfr" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.547822 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-jlkfr" Feb 18 15:34:34 crc kubenswrapper[4968]: I0218 15:34:34.767412 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-jlkfr"] Feb 18 15:34:35 crc kubenswrapper[4968]: I0218 15:34:35.096801 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-jlkfr" event={"ID":"63e09cfe-38d6-44fd-859c-3bc876d8808f","Type":"ContainerStarted","Data":"ff85cfa1b97a74897f0434628bf09e4513d07860103cbf61b340204d69f04a3b"} Feb 18 15:34:35 crc kubenswrapper[4968]: I0218 15:34:35.929263 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:35 crc kubenswrapper[4968]: I0218 15:34:35.933952 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:37 crc kubenswrapper[4968]: I0218 15:34:37.011644 4968 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cwp6r" podUID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" containerName="registry-server" probeResult="failure" output=< Feb 18 15:34:37 crc kubenswrapper[4968]: timeout: failed to connect service ":50051" within 1s Feb 18 15:34:37 crc kubenswrapper[4968]: > Feb 18 15:34:39 crc kubenswrapper[4968]: I0218 15:34:39.123796 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-jlkfr" event={"ID":"63e09cfe-38d6-44fd-859c-3bc876d8808f","Type":"ContainerStarted","Data":"512c7832be9921f4e36b9272f599fffd39f0abd5c99ed6e6654148fa1e0f7d8b"} Feb 18 15:34:39 crc kubenswrapper[4968]: I0218 15:34:39.144861 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:39 crc kubenswrapper[4968]: I0218 15:34:39.144929 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:39 crc kubenswrapper[4968]: I0218 15:34:39.148145 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-694c9596b7-jlkfr" podStartSLOduration=1.666825157 podStartE2EDuration="5.148121819s" podCreationTimestamp="2026-02-18 15:34:34 +0000 UTC" firstStartedPulling="2026-02-18 15:34:34.787841166 +0000 UTC m=+774.173286028" lastFinishedPulling="2026-02-18 15:34:38.269137818 +0000 UTC m=+777.654582690" observedRunningTime="2026-02-18 15:34:39.145183257 +0000 UTC m=+778.530628139" watchObservedRunningTime="2026-02-18 15:34:39.148121819 +0000 UTC m=+778.533566681" Feb 18 15:34:39 crc kubenswrapper[4968]: I0218 15:34:39.188231 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:40 crc kubenswrapper[4968]: I0218 15:34:40.173611 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:41 crc kubenswrapper[4968]: I0218 15:34:41.409386 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c82lb"] Feb 18 15:34:42 crc kubenswrapper[4968]: I0218 15:34:42.145797 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c82lb" podUID="d05ee3b8-c378-467d-a0cd-d58121ae17a9" containerName="registry-server" containerID="cri-o://b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80" gracePeriod=2 Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.073085 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.152181 4968 generic.go:334] "Generic (PLEG): container finished" podID="d05ee3b8-c378-467d-a0cd-d58121ae17a9" containerID="b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80" exitCode=0 Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.152344 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c82lb" event={"ID":"d05ee3b8-c378-467d-a0cd-d58121ae17a9","Type":"ContainerDied","Data":"b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80"} Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.152468 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c82lb" event={"ID":"d05ee3b8-c378-467d-a0cd-d58121ae17a9","Type":"ContainerDied","Data":"0c39aa692a2a7e6ba8530203a715607b50232888a22b231d78fc15cf53843f7d"} Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.152414 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c82lb" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.152493 4968 scope.go:117] "RemoveContainer" containerID="b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.169478 4968 scope.go:117] "RemoveContainer" containerID="dc2d73943abae31e1b7072dcde3a97504a66056363d924193efd0b4ce8e3fe81" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.183022 4968 scope.go:117] "RemoveContainer" containerID="3c88df8ac907236c1332e38a2c8cd86f0a5d05dedc0b91bea650e06e51e78ecc" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.210506 4968 scope.go:117] "RemoveContainer" containerID="b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80" Feb 18 15:34:43 crc kubenswrapper[4968]: E0218 15:34:43.210989 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80\": container with ID starting with b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80 not found: ID does not exist" containerID="b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.211017 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80"} err="failed to get container status \"b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80\": rpc error: code = NotFound desc = could not find container \"b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80\": container with ID starting with b589c2d5832cb478d62d15d5ac5214b7d8f702a44aef48b15c6c62f470613d80 not found: ID does not exist" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.211039 4968 scope.go:117] "RemoveContainer" containerID="dc2d73943abae31e1b7072dcde3a97504a66056363d924193efd0b4ce8e3fe81" Feb 18 15:34:43 crc kubenswrapper[4968]: E0218 15:34:43.211317 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc2d73943abae31e1b7072dcde3a97504a66056363d924193efd0b4ce8e3fe81\": container with ID starting with dc2d73943abae31e1b7072dcde3a97504a66056363d924193efd0b4ce8e3fe81 not found: ID does not exist" containerID="dc2d73943abae31e1b7072dcde3a97504a66056363d924193efd0b4ce8e3fe81" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.211339 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc2d73943abae31e1b7072dcde3a97504a66056363d924193efd0b4ce8e3fe81"} err="failed to get container status \"dc2d73943abae31e1b7072dcde3a97504a66056363d924193efd0b4ce8e3fe81\": rpc error: code = NotFound desc = could not find container \"dc2d73943abae31e1b7072dcde3a97504a66056363d924193efd0b4ce8e3fe81\": container with ID starting with dc2d73943abae31e1b7072dcde3a97504a66056363d924193efd0b4ce8e3fe81 not found: ID does not exist" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.211353 4968 scope.go:117] "RemoveContainer" containerID="3c88df8ac907236c1332e38a2c8cd86f0a5d05dedc0b91bea650e06e51e78ecc" Feb 18 15:34:43 crc kubenswrapper[4968]: E0218 15:34:43.211759 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c88df8ac907236c1332e38a2c8cd86f0a5d05dedc0b91bea650e06e51e78ecc\": container with ID starting with 3c88df8ac907236c1332e38a2c8cd86f0a5d05dedc0b91bea650e06e51e78ecc not found: ID does not exist" containerID="3c88df8ac907236c1332e38a2c8cd86f0a5d05dedc0b91bea650e06e51e78ecc" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.211786 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c88df8ac907236c1332e38a2c8cd86f0a5d05dedc0b91bea650e06e51e78ecc"} err="failed to get container status \"3c88df8ac907236c1332e38a2c8cd86f0a5d05dedc0b91bea650e06e51e78ecc\": rpc error: code = NotFound desc = could not find container \"3c88df8ac907236c1332e38a2c8cd86f0a5d05dedc0b91bea650e06e51e78ecc\": container with ID starting with 3c88df8ac907236c1332e38a2c8cd86f0a5d05dedc0b91bea650e06e51e78ecc not found: ID does not exist" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.229310 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnbwc\" (UniqueName: \"kubernetes.io/projected/d05ee3b8-c378-467d-a0cd-d58121ae17a9-kube-api-access-fnbwc\") pod \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\" (UID: \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\") " Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.229397 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05ee3b8-c378-467d-a0cd-d58121ae17a9-catalog-content\") pod \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\" (UID: \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\") " Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.229419 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05ee3b8-c378-467d-a0cd-d58121ae17a9-utilities\") pod \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\" (UID: \"d05ee3b8-c378-467d-a0cd-d58121ae17a9\") " Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.230664 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05ee3b8-c378-467d-a0cd-d58121ae17a9-utilities" (OuterVolumeSpecName: "utilities") pod "d05ee3b8-c378-467d-a0cd-d58121ae17a9" (UID: "d05ee3b8-c378-467d-a0cd-d58121ae17a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.234807 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d05ee3b8-c378-467d-a0cd-d58121ae17a9-kube-api-access-fnbwc" (OuterVolumeSpecName: "kube-api-access-fnbwc") pod "d05ee3b8-c378-467d-a0cd-d58121ae17a9" (UID: "d05ee3b8-c378-467d-a0cd-d58121ae17a9"). InnerVolumeSpecName "kube-api-access-fnbwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.252826 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05ee3b8-c378-467d-a0cd-d58121ae17a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d05ee3b8-c378-467d-a0cd-d58121ae17a9" (UID: "d05ee3b8-c378-467d-a0cd-d58121ae17a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.332316 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnbwc\" (UniqueName: \"kubernetes.io/projected/d05ee3b8-c378-467d-a0cd-d58121ae17a9-kube-api-access-fnbwc\") on node \"crc\" DevicePath \"\"" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.332344 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05ee3b8-c378-467d-a0cd-d58121ae17a9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.332356 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05ee3b8-c378-467d-a0cd-d58121ae17a9-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.498696 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c82lb"] Feb 18 15:34:43 crc kubenswrapper[4968]: I0218 15:34:43.507603 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c82lb"] Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.343617 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-cfdkm"] Feb 18 15:34:44 crc kubenswrapper[4968]: E0218 15:34:44.343951 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05ee3b8-c378-467d-a0cd-d58121ae17a9" containerName="registry-server" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.343970 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05ee3b8-c378-467d-a0cd-d58121ae17a9" containerName="registry-server" Feb 18 15:34:44 crc kubenswrapper[4968]: E0218 15:34:44.343994 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05ee3b8-c378-467d-a0cd-d58121ae17a9" containerName="extract-utilities" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.344005 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05ee3b8-c378-467d-a0cd-d58121ae17a9" containerName="extract-utilities" Feb 18 15:34:44 crc kubenswrapper[4968]: E0218 15:34:44.344029 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05ee3b8-c378-467d-a0cd-d58121ae17a9" containerName="extract-content" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.344042 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05ee3b8-c378-467d-a0cd-d58121ae17a9" containerName="extract-content" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.344204 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05ee3b8-c378-467d-a0cd-d58121ae17a9" containerName="registry-server" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.345065 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-cfdkm" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.347285 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-htzjr" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.349208 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb"] Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.350090 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.357478 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.367668 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.367736 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.367814 4968 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.368524 4968 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"47c8c50ad8111d46dcdc1997d2e8171ff7419555273a45fc7043517f33afc7b7"} pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.368608 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" containerID="cri-o://47c8c50ad8111d46dcdc1997d2e8171ff7419555273a45fc7043517f33afc7b7" gracePeriod=600 Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.381140 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-wsb68"] Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.381961 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.392972 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb"] Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.396713 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-cfdkm"] Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.445473 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2wx6\" (UniqueName: \"kubernetes.io/projected/e460ba68-eeb3-4979-a32d-8c2866dfbdb9-kube-api-access-x2wx6\") pod \"nmstate-metrics-58c85c668d-cfdkm\" (UID: \"e460ba68-eeb3-4979-a32d-8c2866dfbdb9\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-cfdkm" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.445556 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/81a82099-84e4-40c8-8dd0-b8803001bebf-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-zwksb\" (UID: \"81a82099-84e4-40c8-8dd0-b8803001bebf\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.445603 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9fpc\" (UniqueName: \"kubernetes.io/projected/81a82099-84e4-40c8-8dd0-b8803001bebf-kube-api-access-k9fpc\") pod \"nmstate-webhook-866bcb46dc-zwksb\" (UID: \"81a82099-84e4-40c8-8dd0-b8803001bebf\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.498239 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd"] Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.499118 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.508526 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.508768 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-gh8vz" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.508935 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.517720 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd"] Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.546977 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcgb7\" (UniqueName: \"kubernetes.io/projected/fbb181f8-e6e8-48c4-9598-30622c5e04d8-kube-api-access-mcgb7\") pod \"nmstate-handler-wsb68\" (UID: \"fbb181f8-e6e8-48c4-9598-30622c5e04d8\") " pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.547032 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/81a82099-84e4-40c8-8dd0-b8803001bebf-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-zwksb\" (UID: \"81a82099-84e4-40c8-8dd0-b8803001bebf\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.547076 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/fbb181f8-e6e8-48c4-9598-30622c5e04d8-nmstate-lock\") pod \"nmstate-handler-wsb68\" (UID: \"fbb181f8-e6e8-48c4-9598-30622c5e04d8\") " pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.547115 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9fpc\" (UniqueName: \"kubernetes.io/projected/81a82099-84e4-40c8-8dd0-b8803001bebf-kube-api-access-k9fpc\") pod \"nmstate-webhook-866bcb46dc-zwksb\" (UID: \"81a82099-84e4-40c8-8dd0-b8803001bebf\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.547141 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/fbb181f8-e6e8-48c4-9598-30622c5e04d8-dbus-socket\") pod \"nmstate-handler-wsb68\" (UID: \"fbb181f8-e6e8-48c4-9598-30622c5e04d8\") " pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.547166 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/fbb181f8-e6e8-48c4-9598-30622c5e04d8-ovs-socket\") pod \"nmstate-handler-wsb68\" (UID: \"fbb181f8-e6e8-48c4-9598-30622c5e04d8\") " pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.547212 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2wx6\" (UniqueName: \"kubernetes.io/projected/e460ba68-eeb3-4979-a32d-8c2866dfbdb9-kube-api-access-x2wx6\") pod \"nmstate-metrics-58c85c668d-cfdkm\" (UID: \"e460ba68-eeb3-4979-a32d-8c2866dfbdb9\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-cfdkm" Feb 18 15:34:44 crc kubenswrapper[4968]: E0218 15:34:44.549931 4968 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Feb 18 15:34:44 crc kubenswrapper[4968]: E0218 15:34:44.550016 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81a82099-84e4-40c8-8dd0-b8803001bebf-tls-key-pair podName:81a82099-84e4-40c8-8dd0-b8803001bebf nodeName:}" failed. No retries permitted until 2026-02-18 15:34:45.049995462 +0000 UTC m=+784.435440324 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/81a82099-84e4-40c8-8dd0-b8803001bebf-tls-key-pair") pod "nmstate-webhook-866bcb46dc-zwksb" (UID: "81a82099-84e4-40c8-8dd0-b8803001bebf") : secret "openshift-nmstate-webhook" not found Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.578056 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2wx6\" (UniqueName: \"kubernetes.io/projected/e460ba68-eeb3-4979-a32d-8c2866dfbdb9-kube-api-access-x2wx6\") pod \"nmstate-metrics-58c85c668d-cfdkm\" (UID: \"e460ba68-eeb3-4979-a32d-8c2866dfbdb9\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-cfdkm" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.584975 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9fpc\" (UniqueName: \"kubernetes.io/projected/81a82099-84e4-40c8-8dd0-b8803001bebf-kube-api-access-k9fpc\") pod \"nmstate-webhook-866bcb46dc-zwksb\" (UID: \"81a82099-84e4-40c8-8dd0-b8803001bebf\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.661351 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-cfdkm" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.661709 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/fbb181f8-e6e8-48c4-9598-30622c5e04d8-nmstate-lock\") pod \"nmstate-handler-wsb68\" (UID: \"fbb181f8-e6e8-48c4-9598-30622c5e04d8\") " pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.661798 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/fbb181f8-e6e8-48c4-9598-30622c5e04d8-dbus-socket\") pod \"nmstate-handler-wsb68\" (UID: \"fbb181f8-e6e8-48c4-9598-30622c5e04d8\") " pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.661815 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/fbb181f8-e6e8-48c4-9598-30622c5e04d8-nmstate-lock\") pod \"nmstate-handler-wsb68\" (UID: \"fbb181f8-e6e8-48c4-9598-30622c5e04d8\") " pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.661876 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp9b2\" (UniqueName: \"kubernetes.io/projected/3af661c9-d7d5-4b18-8b4a-d6dac9565b6c-kube-api-access-tp9b2\") pod \"nmstate-console-plugin-5c78fc5d65-prmzd\" (UID: \"3af661c9-d7d5-4b18-8b4a-d6dac9565b6c\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.661901 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/fbb181f8-e6e8-48c4-9598-30622c5e04d8-ovs-socket\") pod \"nmstate-handler-wsb68\" (UID: \"fbb181f8-e6e8-48c4-9598-30622c5e04d8\") " pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.661920 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3af661c9-d7d5-4b18-8b4a-d6dac9565b6c-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-prmzd\" (UID: \"3af661c9-d7d5-4b18-8b4a-d6dac9565b6c\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.661980 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcgb7\" (UniqueName: \"kubernetes.io/projected/fbb181f8-e6e8-48c4-9598-30622c5e04d8-kube-api-access-mcgb7\") pod \"nmstate-handler-wsb68\" (UID: \"fbb181f8-e6e8-48c4-9598-30622c5e04d8\") " pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.662021 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3af661c9-d7d5-4b18-8b4a-d6dac9565b6c-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-prmzd\" (UID: \"3af661c9-d7d5-4b18-8b4a-d6dac9565b6c\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.662038 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/fbb181f8-e6e8-48c4-9598-30622c5e04d8-ovs-socket\") pod \"nmstate-handler-wsb68\" (UID: \"fbb181f8-e6e8-48c4-9598-30622c5e04d8\") " pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.662186 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/fbb181f8-e6e8-48c4-9598-30622c5e04d8-dbus-socket\") pod \"nmstate-handler-wsb68\" (UID: \"fbb181f8-e6e8-48c4-9598-30622c5e04d8\") " pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.690066 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-546898654f-mx5s9"] Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.690793 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.691905 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcgb7\" (UniqueName: \"kubernetes.io/projected/fbb181f8-e6e8-48c4-9598-30622c5e04d8-kube-api-access-mcgb7\") pod \"nmstate-handler-wsb68\" (UID: \"fbb181f8-e6e8-48c4-9598-30622c5e04d8\") " pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.705609 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-546898654f-mx5s9"] Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.714176 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.764685 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3af661c9-d7d5-4b18-8b4a-d6dac9565b6c-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-prmzd\" (UID: \"3af661c9-d7d5-4b18-8b4a-d6dac9565b6c\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.764917 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp9b2\" (UniqueName: \"kubernetes.io/projected/3af661c9-d7d5-4b18-8b4a-d6dac9565b6c-kube-api-access-tp9b2\") pod \"nmstate-console-plugin-5c78fc5d65-prmzd\" (UID: \"3af661c9-d7d5-4b18-8b4a-d6dac9565b6c\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.764952 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3af661c9-d7d5-4b18-8b4a-d6dac9565b6c-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-prmzd\" (UID: \"3af661c9-d7d5-4b18-8b4a-d6dac9565b6c\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" Feb 18 15:34:44 crc kubenswrapper[4968]: E0218 15:34:44.765135 4968 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Feb 18 15:34:44 crc kubenswrapper[4968]: E0218 15:34:44.765270 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3af661c9-d7d5-4b18-8b4a-d6dac9565b6c-plugin-serving-cert podName:3af661c9-d7d5-4b18-8b4a-d6dac9565b6c nodeName:}" failed. No retries permitted until 2026-02-18 15:34:45.265255934 +0000 UTC m=+784.650700796 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/3af661c9-d7d5-4b18-8b4a-d6dac9565b6c-plugin-serving-cert") pod "nmstate-console-plugin-5c78fc5d65-prmzd" (UID: "3af661c9-d7d5-4b18-8b4a-d6dac9565b6c") : secret "plugin-serving-cert" not found Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.765798 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3af661c9-d7d5-4b18-8b4a-d6dac9565b6c-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-prmzd\" (UID: \"3af661c9-d7d5-4b18-8b4a-d6dac9565b6c\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.780988 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp9b2\" (UniqueName: \"kubernetes.io/projected/3af661c9-d7d5-4b18-8b4a-d6dac9565b6c-kube-api-access-tp9b2\") pod \"nmstate-console-plugin-5c78fc5d65-prmzd\" (UID: \"3af661c9-d7d5-4b18-8b4a-d6dac9565b6c\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.865794 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1355be3b-1339-4c83-862b-a4d5fef67500-console-oauth-config\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.865830 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1355be3b-1339-4c83-862b-a4d5fef67500-trusted-ca-bundle\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.865940 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1355be3b-1339-4c83-862b-a4d5fef67500-console-serving-cert\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.865975 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1355be3b-1339-4c83-862b-a4d5fef67500-console-config\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.865992 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1355be3b-1339-4c83-862b-a4d5fef67500-oauth-serving-cert\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.866007 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttjp9\" (UniqueName: \"kubernetes.io/projected/1355be3b-1339-4c83-862b-a4d5fef67500-kube-api-access-ttjp9\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.866057 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1355be3b-1339-4c83-862b-a4d5fef67500-service-ca\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.906039 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-cfdkm"] Feb 18 15:34:44 crc kubenswrapper[4968]: W0218 15:34:44.913660 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode460ba68_eeb3_4979_a32d_8c2866dfbdb9.slice/crio-cf4835ac89bf01fd41b53803dd72bf1e11c8274d450b16f57da4ff164c4d7874 WatchSource:0}: Error finding container cf4835ac89bf01fd41b53803dd72bf1e11c8274d450b16f57da4ff164c4d7874: Status 404 returned error can't find the container with id cf4835ac89bf01fd41b53803dd72bf1e11c8274d450b16f57da4ff164c4d7874 Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.967125 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1355be3b-1339-4c83-862b-a4d5fef67500-console-config\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.967188 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1355be3b-1339-4c83-862b-a4d5fef67500-oauth-serving-cert\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.967221 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttjp9\" (UniqueName: \"kubernetes.io/projected/1355be3b-1339-4c83-862b-a4d5fef67500-kube-api-access-ttjp9\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.967297 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1355be3b-1339-4c83-862b-a4d5fef67500-service-ca\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.967369 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1355be3b-1339-4c83-862b-a4d5fef67500-console-oauth-config\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.967403 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1355be3b-1339-4c83-862b-a4d5fef67500-trusted-ca-bundle\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.968308 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1355be3b-1339-4c83-862b-a4d5fef67500-console-config\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.968341 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1355be3b-1339-4c83-862b-a4d5fef67500-oauth-serving-cert\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.968614 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1355be3b-1339-4c83-862b-a4d5fef67500-service-ca\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.969384 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1355be3b-1339-4c83-862b-a4d5fef67500-trusted-ca-bundle\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.969445 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1355be3b-1339-4c83-862b-a4d5fef67500-console-serving-cert\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.971133 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1355be3b-1339-4c83-862b-a4d5fef67500-console-serving-cert\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.971285 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1355be3b-1339-4c83-862b-a4d5fef67500-console-oauth-config\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:44 crc kubenswrapper[4968]: I0218 15:34:44.983733 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttjp9\" (UniqueName: \"kubernetes.io/projected/1355be3b-1339-4c83-862b-a4d5fef67500-kube-api-access-ttjp9\") pod \"console-546898654f-mx5s9\" (UID: \"1355be3b-1339-4c83-862b-a4d5fef67500\") " pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.054597 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.070896 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/81a82099-84e4-40c8-8dd0-b8803001bebf-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-zwksb\" (UID: \"81a82099-84e4-40c8-8dd0-b8803001bebf\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.076034 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/81a82099-84e4-40c8-8dd0-b8803001bebf-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-zwksb\" (UID: \"81a82099-84e4-40c8-8dd0-b8803001bebf\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.168765 4968 generic.go:334] "Generic (PLEG): container finished" podID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerID="47c8c50ad8111d46dcdc1997d2e8171ff7419555273a45fc7043517f33afc7b7" exitCode=0 Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.168786 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerDied","Data":"47c8c50ad8111d46dcdc1997d2e8171ff7419555273a45fc7043517f33afc7b7"} Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.169322 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerStarted","Data":"2024fbd9a3a558a0c7ebda43f30fc1d9386f6d00027774f8f91dd895439fc899"} Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.169353 4968 scope.go:117] "RemoveContainer" containerID="c21daacf9a62d32cfa427d33ebe792f5f91bea8de4e6e2738e79b36fa1f3f589" Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.172786 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-cfdkm" event={"ID":"e460ba68-eeb3-4979-a32d-8c2866dfbdb9","Type":"ContainerStarted","Data":"cf4835ac89bf01fd41b53803dd72bf1e11c8274d450b16f57da4ff164c4d7874"} Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.175830 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-wsb68" event={"ID":"fbb181f8-e6e8-48c4-9598-30622c5e04d8","Type":"ContainerStarted","Data":"65af12e4d3ada259d0eef41380ef1acd480c083713abb0cfaa854f46f413d0da"} Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.243803 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d05ee3b8-c378-467d-a0cd-d58121ae17a9" path="/var/lib/kubelet/pods/d05ee3b8-c378-467d-a0cd-d58121ae17a9/volumes" Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.270067 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.273246 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3af661c9-d7d5-4b18-8b4a-d6dac9565b6c-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-prmzd\" (UID: \"3af661c9-d7d5-4b18-8b4a-d6dac9565b6c\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.282093 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3af661c9-d7d5-4b18-8b4a-d6dac9565b6c-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-prmzd\" (UID: \"3af661c9-d7d5-4b18-8b4a-d6dac9565b6c\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.417530 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.478535 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb"] Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.506362 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-546898654f-mx5s9"] Feb 18 15:34:45 crc kubenswrapper[4968]: W0218 15:34:45.560028 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1355be3b_1339_4c83_862b_a4d5fef67500.slice/crio-e327636bcc40235ceee6439f9be585223139c75fb73ed2d17c757a9f2aa63b67 WatchSource:0}: Error finding container e327636bcc40235ceee6439f9be585223139c75fb73ed2d17c757a9f2aa63b67: Status 404 returned error can't find the container with id e327636bcc40235ceee6439f9be585223139c75fb73ed2d17c757a9f2aa63b67 Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.722834 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd"] Feb 18 15:34:45 crc kubenswrapper[4968]: I0218 15:34:45.980489 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:46 crc kubenswrapper[4968]: I0218 15:34:46.020314 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:46 crc kubenswrapper[4968]: I0218 15:34:46.199100 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" event={"ID":"3af661c9-d7d5-4b18-8b4a-d6dac9565b6c","Type":"ContainerStarted","Data":"d2cdecffd6bbc7a5a299f293d306fc3b2a0c1def872ecd7b24e7ca3c00923637"} Feb 18 15:34:46 crc kubenswrapper[4968]: I0218 15:34:46.200408 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-546898654f-mx5s9" event={"ID":"1355be3b-1339-4c83-862b-a4d5fef67500","Type":"ContainerStarted","Data":"5bb45c415b7bbae2b560e1056370a88d1fbd01a20433dee844ee966f62dccdeb"} Feb 18 15:34:46 crc kubenswrapper[4968]: I0218 15:34:46.200455 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-546898654f-mx5s9" event={"ID":"1355be3b-1339-4c83-862b-a4d5fef67500","Type":"ContainerStarted","Data":"e327636bcc40235ceee6439f9be585223139c75fb73ed2d17c757a9f2aa63b67"} Feb 18 15:34:46 crc kubenswrapper[4968]: I0218 15:34:46.204945 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" event={"ID":"81a82099-84e4-40c8-8dd0-b8803001bebf","Type":"ContainerStarted","Data":"1e9e85dde468e800f15b9d1e10129132f3622178f32159aaf99281c27b8ab54c"} Feb 18 15:34:46 crc kubenswrapper[4968]: I0218 15:34:46.228753 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-546898654f-mx5s9" podStartSLOduration=2.228665655 podStartE2EDuration="2.228665655s" podCreationTimestamp="2026-02-18 15:34:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:34:46.222023609 +0000 UTC m=+785.607468511" watchObservedRunningTime="2026-02-18 15:34:46.228665655 +0000 UTC m=+785.614110507" Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.207809 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wzbm2"] Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.209274 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.245300 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wzbm2"] Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.308386 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caaee238-b2f2-41b9-bd65-47fce5f5a77f-catalog-content\") pod \"certified-operators-wzbm2\" (UID: \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\") " pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.308463 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caaee238-b2f2-41b9-bd65-47fce5f5a77f-utilities\") pod \"certified-operators-wzbm2\" (UID: \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\") " pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.308488 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh5rk\" (UniqueName: \"kubernetes.io/projected/caaee238-b2f2-41b9-bd65-47fce5f5a77f-kube-api-access-zh5rk\") pod \"certified-operators-wzbm2\" (UID: \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\") " pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.412481 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caaee238-b2f2-41b9-bd65-47fce5f5a77f-catalog-content\") pod \"certified-operators-wzbm2\" (UID: \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\") " pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.413115 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caaee238-b2f2-41b9-bd65-47fce5f5a77f-catalog-content\") pod \"certified-operators-wzbm2\" (UID: \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\") " pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.414142 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caaee238-b2f2-41b9-bd65-47fce5f5a77f-utilities\") pod \"certified-operators-wzbm2\" (UID: \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\") " pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.414192 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh5rk\" (UniqueName: \"kubernetes.io/projected/caaee238-b2f2-41b9-bd65-47fce5f5a77f-kube-api-access-zh5rk\") pod \"certified-operators-wzbm2\" (UID: \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\") " pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.417683 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caaee238-b2f2-41b9-bd65-47fce5f5a77f-utilities\") pod \"certified-operators-wzbm2\" (UID: \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\") " pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.455159 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh5rk\" (UniqueName: \"kubernetes.io/projected/caaee238-b2f2-41b9-bd65-47fce5f5a77f-kube-api-access-zh5rk\") pod \"certified-operators-wzbm2\" (UID: \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\") " pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.544072 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:47 crc kubenswrapper[4968]: I0218 15:34:47.974308 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wzbm2"] Feb 18 15:34:48 crc kubenswrapper[4968]: I0218 15:34:48.231715 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-cfdkm" event={"ID":"e460ba68-eeb3-4979-a32d-8c2866dfbdb9","Type":"ContainerStarted","Data":"6a583ec3d7415a3840a3bd123a3282935c9088c984d3d087a478e3a0ff929f77"} Feb 18 15:34:48 crc kubenswrapper[4968]: I0218 15:34:48.233231 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" event={"ID":"81a82099-84e4-40c8-8dd0-b8803001bebf","Type":"ContainerStarted","Data":"6a0e47297561d5657db05ab075e6b8323f81f072b2f35ae7476dcc94206ec7e2"} Feb 18 15:34:48 crc kubenswrapper[4968]: I0218 15:34:48.233365 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" Feb 18 15:34:48 crc kubenswrapper[4968]: I0218 15:34:48.234520 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-wsb68" event={"ID":"fbb181f8-e6e8-48c4-9598-30622c5e04d8","Type":"ContainerStarted","Data":"4e9a74e1e05b7b8e53c9d2ea61cea9772ce7f44925f7e69e8e6c12dc9898bd69"} Feb 18 15:34:48 crc kubenswrapper[4968]: I0218 15:34:48.234623 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:48 crc kubenswrapper[4968]: I0218 15:34:48.236292 4968 generic.go:334] "Generic (PLEG): container finished" podID="caaee238-b2f2-41b9-bd65-47fce5f5a77f" containerID="802d3815c0c34a82bf719e896e4b4123ecce4fecd2e270927e0f2506fe7da48c" exitCode=0 Feb 18 15:34:48 crc kubenswrapper[4968]: I0218 15:34:48.236328 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzbm2" event={"ID":"caaee238-b2f2-41b9-bd65-47fce5f5a77f","Type":"ContainerDied","Data":"802d3815c0c34a82bf719e896e4b4123ecce4fecd2e270927e0f2506fe7da48c"} Feb 18 15:34:48 crc kubenswrapper[4968]: I0218 15:34:48.236357 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzbm2" event={"ID":"caaee238-b2f2-41b9-bd65-47fce5f5a77f","Type":"ContainerStarted","Data":"d71779f4fdd842b3a3fb6e7fde8660d586cdd77ef3c213271b0fb5004e5b6e8f"} Feb 18 15:34:48 crc kubenswrapper[4968]: I0218 15:34:48.252963 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" podStartSLOduration=1.922149369 podStartE2EDuration="4.252949665s" podCreationTimestamp="2026-02-18 15:34:44 +0000 UTC" firstStartedPulling="2026-02-18 15:34:45.50682415 +0000 UTC m=+784.892269012" lastFinishedPulling="2026-02-18 15:34:47.837624446 +0000 UTC m=+787.223069308" observedRunningTime="2026-02-18 15:34:48.248143081 +0000 UTC m=+787.633587963" watchObservedRunningTime="2026-02-18 15:34:48.252949665 +0000 UTC m=+787.638394527" Feb 18 15:34:48 crc kubenswrapper[4968]: I0218 15:34:48.287188 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-wsb68" podStartSLOduration=1.227623398 podStartE2EDuration="4.287168212s" podCreationTimestamp="2026-02-18 15:34:44 +0000 UTC" firstStartedPulling="2026-02-18 15:34:44.765420188 +0000 UTC m=+784.150865050" lastFinishedPulling="2026-02-18 15:34:47.824965002 +0000 UTC m=+787.210409864" observedRunningTime="2026-02-18 15:34:48.28566924 +0000 UTC m=+787.671114102" watchObservedRunningTime="2026-02-18 15:34:48.287168212 +0000 UTC m=+787.672613074" Feb 18 15:34:49 crc kubenswrapper[4968]: I0218 15:34:49.243484 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" event={"ID":"3af661c9-d7d5-4b18-8b4a-d6dac9565b6c","Type":"ContainerStarted","Data":"6fb1059478215fd575cf0db4d2eda831b26d0615d2ffe5f11b474a06ca7e0628"} Feb 18 15:34:49 crc kubenswrapper[4968]: I0218 15:34:49.246294 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzbm2" event={"ID":"caaee238-b2f2-41b9-bd65-47fce5f5a77f","Type":"ContainerStarted","Data":"ce2747e7f365d1b02f1492d9eae75a84d7498f10e5e869cfbcf52b430883990c"} Feb 18 15:34:49 crc kubenswrapper[4968]: I0218 15:34:49.280717 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-prmzd" podStartSLOduration=2.1159488 podStartE2EDuration="5.280694607s" podCreationTimestamp="2026-02-18 15:34:44 +0000 UTC" firstStartedPulling="2026-02-18 15:34:45.733702867 +0000 UTC m=+785.119147729" lastFinishedPulling="2026-02-18 15:34:48.898448684 +0000 UTC m=+788.283893536" observedRunningTime="2026-02-18 15:34:49.259116233 +0000 UTC m=+788.644561095" watchObservedRunningTime="2026-02-18 15:34:49.280694607 +0000 UTC m=+788.666139469" Feb 18 15:34:49 crc kubenswrapper[4968]: I0218 15:34:49.806695 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cwp6r"] Feb 18 15:34:49 crc kubenswrapper[4968]: I0218 15:34:49.807726 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cwp6r" podUID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" containerName="registry-server" containerID="cri-o://4d984d9c4c914066fcb166329e1de8982a2ffb666b7f8d69c336aecc6423b52b" gracePeriod=2 Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.262413 4968 generic.go:334] "Generic (PLEG): container finished" podID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" containerID="4d984d9c4c914066fcb166329e1de8982a2ffb666b7f8d69c336aecc6423b52b" exitCode=0 Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.262490 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwp6r" event={"ID":"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642","Type":"ContainerDied","Data":"4d984d9c4c914066fcb166329e1de8982a2ffb666b7f8d69c336aecc6423b52b"} Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.264796 4968 generic.go:334] "Generic (PLEG): container finished" podID="caaee238-b2f2-41b9-bd65-47fce5f5a77f" containerID="ce2747e7f365d1b02f1492d9eae75a84d7498f10e5e869cfbcf52b430883990c" exitCode=0 Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.264887 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzbm2" event={"ID":"caaee238-b2f2-41b9-bd65-47fce5f5a77f","Type":"ContainerDied","Data":"ce2747e7f365d1b02f1492d9eae75a84d7498f10e5e869cfbcf52b430883990c"} Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.336691 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.458496 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwj59\" (UniqueName: \"kubernetes.io/projected/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-kube-api-access-kwj59\") pod \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\" (UID: \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\") " Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.458634 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-catalog-content\") pod \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\" (UID: \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\") " Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.458698 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-utilities\") pod \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\" (UID: \"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642\") " Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.461150 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-utilities" (OuterVolumeSpecName: "utilities") pod "3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" (UID: "3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.467049 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-kube-api-access-kwj59" (OuterVolumeSpecName: "kube-api-access-kwj59") pod "3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" (UID: "3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642"). InnerVolumeSpecName "kube-api-access-kwj59". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.561172 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwj59\" (UniqueName: \"kubernetes.io/projected/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-kube-api-access-kwj59\") on node \"crc\" DevicePath \"\"" Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.561218 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.575354 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" (UID: "3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:34:50 crc kubenswrapper[4968]: I0218 15:34:50.665168 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:34:51 crc kubenswrapper[4968]: I0218 15:34:51.277148 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzbm2" event={"ID":"caaee238-b2f2-41b9-bd65-47fce5f5a77f","Type":"ContainerStarted","Data":"84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab"} Feb 18 15:34:51 crc kubenswrapper[4968]: I0218 15:34:51.279817 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-cfdkm" event={"ID":"e460ba68-eeb3-4979-a32d-8c2866dfbdb9","Type":"ContainerStarted","Data":"bdcd79e8bdf4fb81978f3602ad18cf9fb117b9e368f25276cc692194228da79e"} Feb 18 15:34:51 crc kubenswrapper[4968]: I0218 15:34:51.287119 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwp6r" event={"ID":"3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642","Type":"ContainerDied","Data":"934aaa129af7b3c28fd5ba07a21e2380e7cb7e56bc808db529125a8d07de7111"} Feb 18 15:34:51 crc kubenswrapper[4968]: I0218 15:34:51.287200 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cwp6r" Feb 18 15:34:51 crc kubenswrapper[4968]: I0218 15:34:51.287207 4968 scope.go:117] "RemoveContainer" containerID="4d984d9c4c914066fcb166329e1de8982a2ffb666b7f8d69c336aecc6423b52b" Feb 18 15:34:51 crc kubenswrapper[4968]: I0218 15:34:51.321425 4968 scope.go:117] "RemoveContainer" containerID="8e1baebd50335277fde9216691678ac2d8e03a2f5eecbb8cb58de82bafe4a767" Feb 18 15:34:51 crc kubenswrapper[4968]: I0218 15:34:51.347613 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58c85c668d-cfdkm" podStartSLOduration=2.075253581 podStartE2EDuration="7.347591959s" podCreationTimestamp="2026-02-18 15:34:44 +0000 UTC" firstStartedPulling="2026-02-18 15:34:44.916124145 +0000 UTC m=+784.301569007" lastFinishedPulling="2026-02-18 15:34:50.188462533 +0000 UTC m=+789.573907385" observedRunningTime="2026-02-18 15:34:51.345558852 +0000 UTC m=+790.731003734" watchObservedRunningTime="2026-02-18 15:34:51.347591959 +0000 UTC m=+790.733036831" Feb 18 15:34:51 crc kubenswrapper[4968]: I0218 15:34:51.350716 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wzbm2" podStartSLOduration=1.685853295 podStartE2EDuration="4.350701156s" podCreationTimestamp="2026-02-18 15:34:47 +0000 UTC" firstStartedPulling="2026-02-18 15:34:48.237353189 +0000 UTC m=+787.622798041" lastFinishedPulling="2026-02-18 15:34:50.90220103 +0000 UTC m=+790.287645902" observedRunningTime="2026-02-18 15:34:51.311979613 +0000 UTC m=+790.697424485" watchObservedRunningTime="2026-02-18 15:34:51.350701156 +0000 UTC m=+790.736146028" Feb 18 15:34:51 crc kubenswrapper[4968]: I0218 15:34:51.362100 4968 scope.go:117] "RemoveContainer" containerID="0390185bdf60d40a58dee59b18d551aa98b8f34a590a8a22f95c7bd00666e515" Feb 18 15:34:51 crc kubenswrapper[4968]: I0218 15:34:51.371583 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cwp6r"] Feb 18 15:34:51 crc kubenswrapper[4968]: I0218 15:34:51.376491 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cwp6r"] Feb 18 15:34:53 crc kubenswrapper[4968]: I0218 15:34:53.237593 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" path="/var/lib/kubelet/pods/3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642/volumes" Feb 18 15:34:54 crc kubenswrapper[4968]: I0218 15:34:54.750163 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-wsb68" Feb 18 15:34:55 crc kubenswrapper[4968]: I0218 15:34:55.055643 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:55 crc kubenswrapper[4968]: I0218 15:34:55.055705 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:55 crc kubenswrapper[4968]: I0218 15:34:55.067583 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:55 crc kubenswrapper[4968]: I0218 15:34:55.322930 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-546898654f-mx5s9" Feb 18 15:34:55 crc kubenswrapper[4968]: I0218 15:34:55.420503 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-s6qsc"] Feb 18 15:34:57 crc kubenswrapper[4968]: I0218 15:34:57.544223 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:57 crc kubenswrapper[4968]: I0218 15:34:57.544657 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:57 crc kubenswrapper[4968]: I0218 15:34:57.616986 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:34:58 crc kubenswrapper[4968]: I0218 15:34:58.414562 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:35:00 crc kubenswrapper[4968]: I0218 15:35:00.403306 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wzbm2"] Feb 18 15:35:00 crc kubenswrapper[4968]: I0218 15:35:00.405251 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wzbm2" podUID="caaee238-b2f2-41b9-bd65-47fce5f5a77f" containerName="registry-server" containerID="cri-o://84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab" gracePeriod=2 Feb 18 15:35:00 crc kubenswrapper[4968]: I0218 15:35:00.764073 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:35:00 crc kubenswrapper[4968]: I0218 15:35:00.927524 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caaee238-b2f2-41b9-bd65-47fce5f5a77f-catalog-content\") pod \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\" (UID: \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\") " Feb 18 15:35:00 crc kubenswrapper[4968]: I0218 15:35:00.927668 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zh5rk\" (UniqueName: \"kubernetes.io/projected/caaee238-b2f2-41b9-bd65-47fce5f5a77f-kube-api-access-zh5rk\") pod \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\" (UID: \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\") " Feb 18 15:35:00 crc kubenswrapper[4968]: I0218 15:35:00.927718 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caaee238-b2f2-41b9-bd65-47fce5f5a77f-utilities\") pod \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\" (UID: \"caaee238-b2f2-41b9-bd65-47fce5f5a77f\") " Feb 18 15:35:00 crc kubenswrapper[4968]: I0218 15:35:00.929222 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caaee238-b2f2-41b9-bd65-47fce5f5a77f-utilities" (OuterVolumeSpecName: "utilities") pod "caaee238-b2f2-41b9-bd65-47fce5f5a77f" (UID: "caaee238-b2f2-41b9-bd65-47fce5f5a77f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:35:00 crc kubenswrapper[4968]: I0218 15:35:00.934047 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caaee238-b2f2-41b9-bd65-47fce5f5a77f-kube-api-access-zh5rk" (OuterVolumeSpecName: "kube-api-access-zh5rk") pod "caaee238-b2f2-41b9-bd65-47fce5f5a77f" (UID: "caaee238-b2f2-41b9-bd65-47fce5f5a77f"). InnerVolumeSpecName "kube-api-access-zh5rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.000281 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caaee238-b2f2-41b9-bd65-47fce5f5a77f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "caaee238-b2f2-41b9-bd65-47fce5f5a77f" (UID: "caaee238-b2f2-41b9-bd65-47fce5f5a77f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.029083 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caaee238-b2f2-41b9-bd65-47fce5f5a77f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.029123 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zh5rk\" (UniqueName: \"kubernetes.io/projected/caaee238-b2f2-41b9-bd65-47fce5f5a77f-kube-api-access-zh5rk\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.029139 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caaee238-b2f2-41b9-bd65-47fce5f5a77f-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.372733 4968 generic.go:334] "Generic (PLEG): container finished" podID="caaee238-b2f2-41b9-bd65-47fce5f5a77f" containerID="84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab" exitCode=0 Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.372847 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzbm2" event={"ID":"caaee238-b2f2-41b9-bd65-47fce5f5a77f","Type":"ContainerDied","Data":"84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab"} Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.372948 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzbm2" event={"ID":"caaee238-b2f2-41b9-bd65-47fce5f5a77f","Type":"ContainerDied","Data":"d71779f4fdd842b3a3fb6e7fde8660d586cdd77ef3c213271b0fb5004e5b6e8f"} Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.372994 4968 scope.go:117] "RemoveContainer" containerID="84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.372868 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wzbm2" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.413468 4968 scope.go:117] "RemoveContainer" containerID="ce2747e7f365d1b02f1492d9eae75a84d7498f10e5e869cfbcf52b430883990c" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.418720 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wzbm2"] Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.426494 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wzbm2"] Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.438324 4968 scope.go:117] "RemoveContainer" containerID="802d3815c0c34a82bf719e896e4b4123ecce4fecd2e270927e0f2506fe7da48c" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.475828 4968 scope.go:117] "RemoveContainer" containerID="84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab" Feb 18 15:35:01 crc kubenswrapper[4968]: E0218 15:35:01.476375 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab\": container with ID starting with 84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab not found: ID does not exist" containerID="84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.476426 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab"} err="failed to get container status \"84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab\": rpc error: code = NotFound desc = could not find container \"84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab\": container with ID starting with 84caf660edd57f629448fdb0302baf3f5ee0257415806ca0456de519747b8cab not found: ID does not exist" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.476458 4968 scope.go:117] "RemoveContainer" containerID="ce2747e7f365d1b02f1492d9eae75a84d7498f10e5e869cfbcf52b430883990c" Feb 18 15:35:01 crc kubenswrapper[4968]: E0218 15:35:01.476868 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce2747e7f365d1b02f1492d9eae75a84d7498f10e5e869cfbcf52b430883990c\": container with ID starting with ce2747e7f365d1b02f1492d9eae75a84d7498f10e5e869cfbcf52b430883990c not found: ID does not exist" containerID="ce2747e7f365d1b02f1492d9eae75a84d7498f10e5e869cfbcf52b430883990c" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.476944 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce2747e7f365d1b02f1492d9eae75a84d7498f10e5e869cfbcf52b430883990c"} err="failed to get container status \"ce2747e7f365d1b02f1492d9eae75a84d7498f10e5e869cfbcf52b430883990c\": rpc error: code = NotFound desc = could not find container \"ce2747e7f365d1b02f1492d9eae75a84d7498f10e5e869cfbcf52b430883990c\": container with ID starting with ce2747e7f365d1b02f1492d9eae75a84d7498f10e5e869cfbcf52b430883990c not found: ID does not exist" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.476997 4968 scope.go:117] "RemoveContainer" containerID="802d3815c0c34a82bf719e896e4b4123ecce4fecd2e270927e0f2506fe7da48c" Feb 18 15:35:01 crc kubenswrapper[4968]: E0218 15:35:01.477516 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"802d3815c0c34a82bf719e896e4b4123ecce4fecd2e270927e0f2506fe7da48c\": container with ID starting with 802d3815c0c34a82bf719e896e4b4123ecce4fecd2e270927e0f2506fe7da48c not found: ID does not exist" containerID="802d3815c0c34a82bf719e896e4b4123ecce4fecd2e270927e0f2506fe7da48c" Feb 18 15:35:01 crc kubenswrapper[4968]: I0218 15:35:01.477555 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"802d3815c0c34a82bf719e896e4b4123ecce4fecd2e270927e0f2506fe7da48c"} err="failed to get container status \"802d3815c0c34a82bf719e896e4b4123ecce4fecd2e270927e0f2506fe7da48c\": rpc error: code = NotFound desc = could not find container \"802d3815c0c34a82bf719e896e4b4123ecce4fecd2e270927e0f2506fe7da48c\": container with ID starting with 802d3815c0c34a82bf719e896e4b4123ecce4fecd2e270927e0f2506fe7da48c not found: ID does not exist" Feb 18 15:35:03 crc kubenswrapper[4968]: I0218 15:35:03.243934 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caaee238-b2f2-41b9-bd65-47fce5f5a77f" path="/var/lib/kubelet/pods/caaee238-b2f2-41b9-bd65-47fce5f5a77f/volumes" Feb 18 15:35:05 crc kubenswrapper[4968]: I0218 15:35:05.279179 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-zwksb" Feb 18 15:35:20 crc kubenswrapper[4968]: I0218 15:35:20.486058 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-s6qsc" podUID="2c7da228-335d-4f50-83e6-4ae152ed4874" containerName="console" containerID="cri-o://d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3" gracePeriod=15 Feb 18 15:35:20 crc kubenswrapper[4968]: I0218 15:35:20.926533 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-s6qsc_2c7da228-335d-4f50-83e6-4ae152ed4874/console/0.log" Feb 18 15:35:20 crc kubenswrapper[4968]: I0218 15:35:20.926887 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.025911 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-oauth-serving-cert\") pod \"2c7da228-335d-4f50-83e6-4ae152ed4874\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.026002 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxx4s\" (UniqueName: \"kubernetes.io/projected/2c7da228-335d-4f50-83e6-4ae152ed4874-kube-api-access-dxx4s\") pod \"2c7da228-335d-4f50-83e6-4ae152ed4874\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.026086 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-service-ca\") pod \"2c7da228-335d-4f50-83e6-4ae152ed4874\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.026113 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2c7da228-335d-4f50-83e6-4ae152ed4874-console-oauth-config\") pod \"2c7da228-335d-4f50-83e6-4ae152ed4874\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.026141 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-console-config\") pod \"2c7da228-335d-4f50-83e6-4ae152ed4874\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.026172 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-trusted-ca-bundle\") pod \"2c7da228-335d-4f50-83e6-4ae152ed4874\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.026200 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c7da228-335d-4f50-83e6-4ae152ed4874-console-serving-cert\") pod \"2c7da228-335d-4f50-83e6-4ae152ed4874\" (UID: \"2c7da228-335d-4f50-83e6-4ae152ed4874\") " Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.026899 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "2c7da228-335d-4f50-83e6-4ae152ed4874" (UID: "2c7da228-335d-4f50-83e6-4ae152ed4874"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.026911 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-service-ca" (OuterVolumeSpecName: "service-ca") pod "2c7da228-335d-4f50-83e6-4ae152ed4874" (UID: "2c7da228-335d-4f50-83e6-4ae152ed4874"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.027299 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-console-config" (OuterVolumeSpecName: "console-config") pod "2c7da228-335d-4f50-83e6-4ae152ed4874" (UID: "2c7da228-335d-4f50-83e6-4ae152ed4874"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.027479 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "2c7da228-335d-4f50-83e6-4ae152ed4874" (UID: "2c7da228-335d-4f50-83e6-4ae152ed4874"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.033257 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c7da228-335d-4f50-83e6-4ae152ed4874-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "2c7da228-335d-4f50-83e6-4ae152ed4874" (UID: "2c7da228-335d-4f50-83e6-4ae152ed4874"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.033648 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c7da228-335d-4f50-83e6-4ae152ed4874-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "2c7da228-335d-4f50-83e6-4ae152ed4874" (UID: "2c7da228-335d-4f50-83e6-4ae152ed4874"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.034104 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c7da228-335d-4f50-83e6-4ae152ed4874-kube-api-access-dxx4s" (OuterVolumeSpecName: "kube-api-access-dxx4s") pod "2c7da228-335d-4f50-83e6-4ae152ed4874" (UID: "2c7da228-335d-4f50-83e6-4ae152ed4874"). InnerVolumeSpecName "kube-api-access-dxx4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.127966 4968 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.128011 4968 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2c7da228-335d-4f50-83e6-4ae152ed4874-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.128028 4968 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-console-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.128041 4968 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.128056 4968 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c7da228-335d-4f50-83e6-4ae152ed4874-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.128069 4968 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2c7da228-335d-4f50-83e6-4ae152ed4874-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.128083 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxx4s\" (UniqueName: \"kubernetes.io/projected/2c7da228-335d-4f50-83e6-4ae152ed4874-kube-api-access-dxx4s\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:21 crc kubenswrapper[4968]: E0218 15:35:21.362667 4968 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c7da228_335d_4f50_83e6_4ae152ed4874.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c7da228_335d_4f50_83e6_4ae152ed4874.slice/crio-000bac82fa9da877c36eb7246df400c7db6a60a1846dfbb811a8bfc0cc4516b0\": RecentStats: unable to find data in memory cache]" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.532192 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-s6qsc_2c7da228-335d-4f50-83e6-4ae152ed4874/console/0.log" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.532270 4968 generic.go:334] "Generic (PLEG): container finished" podID="2c7da228-335d-4f50-83e6-4ae152ed4874" containerID="d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3" exitCode=2 Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.532317 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s6qsc" event={"ID":"2c7da228-335d-4f50-83e6-4ae152ed4874","Type":"ContainerDied","Data":"d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3"} Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.532359 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s6qsc" event={"ID":"2c7da228-335d-4f50-83e6-4ae152ed4874","Type":"ContainerDied","Data":"000bac82fa9da877c36eb7246df400c7db6a60a1846dfbb811a8bfc0cc4516b0"} Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.532393 4968 scope.go:117] "RemoveContainer" containerID="d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.532578 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s6qsc" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.557514 4968 scope.go:117] "RemoveContainer" containerID="d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3" Feb 18 15:35:21 crc kubenswrapper[4968]: E0218 15:35:21.558579 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3\": container with ID starting with d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3 not found: ID does not exist" containerID="d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.558640 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3"} err="failed to get container status \"d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3\": rpc error: code = NotFound desc = could not find container \"d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3\": container with ID starting with d44f6b12066f4c63338fdec64d2cbd74fd6ea911eba8f8209b6d7098308110c3 not found: ID does not exist" Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.563568 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-s6qsc"] Feb 18 15:35:21 crc kubenswrapper[4968]: I0218 15:35:21.572165 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-s6qsc"] Feb 18 15:35:23 crc kubenswrapper[4968]: I0218 15:35:23.245113 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c7da228-335d-4f50-83e6-4ae152ed4874" path="/var/lib/kubelet/pods/2c7da228-335d-4f50-83e6-4ae152ed4874/volumes" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.067300 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh"] Feb 18 15:35:26 crc kubenswrapper[4968]: E0218 15:35:26.067929 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caaee238-b2f2-41b9-bd65-47fce5f5a77f" containerName="extract-content" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.067946 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="caaee238-b2f2-41b9-bd65-47fce5f5a77f" containerName="extract-content" Feb 18 15:35:26 crc kubenswrapper[4968]: E0218 15:35:26.067964 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7da228-335d-4f50-83e6-4ae152ed4874" containerName="console" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.067973 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7da228-335d-4f50-83e6-4ae152ed4874" containerName="console" Feb 18 15:35:26 crc kubenswrapper[4968]: E0218 15:35:26.067987 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" containerName="extract-content" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.067995 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" containerName="extract-content" Feb 18 15:35:26 crc kubenswrapper[4968]: E0218 15:35:26.068018 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caaee238-b2f2-41b9-bd65-47fce5f5a77f" containerName="extract-utilities" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.068026 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="caaee238-b2f2-41b9-bd65-47fce5f5a77f" containerName="extract-utilities" Feb 18 15:35:26 crc kubenswrapper[4968]: E0218 15:35:26.068035 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" containerName="registry-server" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.068044 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" containerName="registry-server" Feb 18 15:35:26 crc kubenswrapper[4968]: E0218 15:35:26.068058 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caaee238-b2f2-41b9-bd65-47fce5f5a77f" containerName="registry-server" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.068066 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="caaee238-b2f2-41b9-bd65-47fce5f5a77f" containerName="registry-server" Feb 18 15:35:26 crc kubenswrapper[4968]: E0218 15:35:26.068079 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" containerName="extract-utilities" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.068087 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" containerName="extract-utilities" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.068239 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f6e41a7-c6bb-4dfd-a5d1-3dd13bf1d642" containerName="registry-server" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.068264 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="caaee238-b2f2-41b9-bd65-47fce5f5a77f" containerName="registry-server" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.068278 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c7da228-335d-4f50-83e6-4ae152ed4874" containerName="console" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.069291 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.072225 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.085971 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh"] Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.091224 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnkfd\" (UniqueName: \"kubernetes.io/projected/e034e9a0-5038-41d3-8372-78e611d8b952-kube-api-access-jnkfd\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh\" (UID: \"e034e9a0-5038-41d3-8372-78e611d8b952\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.091352 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e034e9a0-5038-41d3-8372-78e611d8b952-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh\" (UID: \"e034e9a0-5038-41d3-8372-78e611d8b952\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.091437 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e034e9a0-5038-41d3-8372-78e611d8b952-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh\" (UID: \"e034e9a0-5038-41d3-8372-78e611d8b952\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.192306 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e034e9a0-5038-41d3-8372-78e611d8b952-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh\" (UID: \"e034e9a0-5038-41d3-8372-78e611d8b952\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.192405 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e034e9a0-5038-41d3-8372-78e611d8b952-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh\" (UID: \"e034e9a0-5038-41d3-8372-78e611d8b952\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.192445 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnkfd\" (UniqueName: \"kubernetes.io/projected/e034e9a0-5038-41d3-8372-78e611d8b952-kube-api-access-jnkfd\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh\" (UID: \"e034e9a0-5038-41d3-8372-78e611d8b952\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.192925 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e034e9a0-5038-41d3-8372-78e611d8b952-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh\" (UID: \"e034e9a0-5038-41d3-8372-78e611d8b952\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.192974 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e034e9a0-5038-41d3-8372-78e611d8b952-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh\" (UID: \"e034e9a0-5038-41d3-8372-78e611d8b952\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.216833 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnkfd\" (UniqueName: \"kubernetes.io/projected/e034e9a0-5038-41d3-8372-78e611d8b952-kube-api-access-jnkfd\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh\" (UID: \"e034e9a0-5038-41d3-8372-78e611d8b952\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.390849 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:26 crc kubenswrapper[4968]: I0218 15:35:26.850362 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh"] Feb 18 15:35:27 crc kubenswrapper[4968]: I0218 15:35:27.572540 4968 generic.go:334] "Generic (PLEG): container finished" podID="e034e9a0-5038-41d3-8372-78e611d8b952" containerID="5a3ddd2c5a392b1273f7e7c1feafdd8648745cd4acd72e84077d6abff296f8e4" exitCode=0 Feb 18 15:35:27 crc kubenswrapper[4968]: I0218 15:35:27.572607 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" event={"ID":"e034e9a0-5038-41d3-8372-78e611d8b952","Type":"ContainerDied","Data":"5a3ddd2c5a392b1273f7e7c1feafdd8648745cd4acd72e84077d6abff296f8e4"} Feb 18 15:35:27 crc kubenswrapper[4968]: I0218 15:35:27.572857 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" event={"ID":"e034e9a0-5038-41d3-8372-78e611d8b952","Type":"ContainerStarted","Data":"5bfbec8b1cfcb79d06945515cf43d64c13377622cbce53640a8f4b6a3f00d47e"} Feb 18 15:35:29 crc kubenswrapper[4968]: I0218 15:35:29.588159 4968 generic.go:334] "Generic (PLEG): container finished" podID="e034e9a0-5038-41d3-8372-78e611d8b952" containerID="5b24b9ab5fa1d7770947b6ab3c2d23d025f5be6ee99dd41d24ba59d7805e1cf7" exitCode=0 Feb 18 15:35:29 crc kubenswrapper[4968]: I0218 15:35:29.588224 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" event={"ID":"e034e9a0-5038-41d3-8372-78e611d8b952","Type":"ContainerDied","Data":"5b24b9ab5fa1d7770947b6ab3c2d23d025f5be6ee99dd41d24ba59d7805e1cf7"} Feb 18 15:35:30 crc kubenswrapper[4968]: I0218 15:35:30.601428 4968 generic.go:334] "Generic (PLEG): container finished" podID="e034e9a0-5038-41d3-8372-78e611d8b952" containerID="a4ef7419d6f54796b0e1c7d2e71024b87caf099b79f2e2cea13c85c0b57476b5" exitCode=0 Feb 18 15:35:30 crc kubenswrapper[4968]: I0218 15:35:30.601476 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" event={"ID":"e034e9a0-5038-41d3-8372-78e611d8b952","Type":"ContainerDied","Data":"a4ef7419d6f54796b0e1c7d2e71024b87caf099b79f2e2cea13c85c0b57476b5"} Feb 18 15:35:31 crc kubenswrapper[4968]: I0218 15:35:31.866528 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:31 crc kubenswrapper[4968]: I0218 15:35:31.973060 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e034e9a0-5038-41d3-8372-78e611d8b952-bundle\") pod \"e034e9a0-5038-41d3-8372-78e611d8b952\" (UID: \"e034e9a0-5038-41d3-8372-78e611d8b952\") " Feb 18 15:35:31 crc kubenswrapper[4968]: I0218 15:35:31.973414 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnkfd\" (UniqueName: \"kubernetes.io/projected/e034e9a0-5038-41d3-8372-78e611d8b952-kube-api-access-jnkfd\") pod \"e034e9a0-5038-41d3-8372-78e611d8b952\" (UID: \"e034e9a0-5038-41d3-8372-78e611d8b952\") " Feb 18 15:35:31 crc kubenswrapper[4968]: I0218 15:35:31.973583 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e034e9a0-5038-41d3-8372-78e611d8b952-util\") pod \"e034e9a0-5038-41d3-8372-78e611d8b952\" (UID: \"e034e9a0-5038-41d3-8372-78e611d8b952\") " Feb 18 15:35:31 crc kubenswrapper[4968]: I0218 15:35:31.975021 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e034e9a0-5038-41d3-8372-78e611d8b952-bundle" (OuterVolumeSpecName: "bundle") pod "e034e9a0-5038-41d3-8372-78e611d8b952" (UID: "e034e9a0-5038-41d3-8372-78e611d8b952"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:35:31 crc kubenswrapper[4968]: I0218 15:35:31.985009 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e034e9a0-5038-41d3-8372-78e611d8b952-kube-api-access-jnkfd" (OuterVolumeSpecName: "kube-api-access-jnkfd") pod "e034e9a0-5038-41d3-8372-78e611d8b952" (UID: "e034e9a0-5038-41d3-8372-78e611d8b952"). InnerVolumeSpecName "kube-api-access-jnkfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:35:31 crc kubenswrapper[4968]: I0218 15:35:31.995371 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e034e9a0-5038-41d3-8372-78e611d8b952-util" (OuterVolumeSpecName: "util") pod "e034e9a0-5038-41d3-8372-78e611d8b952" (UID: "e034e9a0-5038-41d3-8372-78e611d8b952"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:35:32 crc kubenswrapper[4968]: I0218 15:35:32.077064 4968 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e034e9a0-5038-41d3-8372-78e611d8b952-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:32 crc kubenswrapper[4968]: I0218 15:35:32.077109 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnkfd\" (UniqueName: \"kubernetes.io/projected/e034e9a0-5038-41d3-8372-78e611d8b952-kube-api-access-jnkfd\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:32 crc kubenswrapper[4968]: I0218 15:35:32.077128 4968 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e034e9a0-5038-41d3-8372-78e611d8b952-util\") on node \"crc\" DevicePath \"\"" Feb 18 15:35:32 crc kubenswrapper[4968]: I0218 15:35:32.619601 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" event={"ID":"e034e9a0-5038-41d3-8372-78e611d8b952","Type":"ContainerDied","Data":"5bfbec8b1cfcb79d06945515cf43d64c13377622cbce53640a8f4b6a3f00d47e"} Feb 18 15:35:32 crc kubenswrapper[4968]: I0218 15:35:32.619650 4968 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bfbec8b1cfcb79d06945515cf43d64c13377622cbce53640a8f4b6a3f00d47e" Feb 18 15:35:32 crc kubenswrapper[4968]: I0218 15:35:32.619707 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.160526 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq"] Feb 18 15:35:44 crc kubenswrapper[4968]: E0218 15:35:44.161289 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e034e9a0-5038-41d3-8372-78e611d8b952" containerName="pull" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.161302 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="e034e9a0-5038-41d3-8372-78e611d8b952" containerName="pull" Feb 18 15:35:44 crc kubenswrapper[4968]: E0218 15:35:44.161310 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e034e9a0-5038-41d3-8372-78e611d8b952" containerName="util" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.161316 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="e034e9a0-5038-41d3-8372-78e611d8b952" containerName="util" Feb 18 15:35:44 crc kubenswrapper[4968]: E0218 15:35:44.161326 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e034e9a0-5038-41d3-8372-78e611d8b952" containerName="extract" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.161333 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="e034e9a0-5038-41d3-8372-78e611d8b952" containerName="extract" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.161436 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="e034e9a0-5038-41d3-8372-78e611d8b952" containerName="extract" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.161826 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.164207 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.164341 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.165126 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.166105 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-fssf8" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.166120 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.178862 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq"] Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.249009 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx4ms\" (UniqueName: \"kubernetes.io/projected/4e91721f-0e20-4dfb-9e17-45107d5cf8ce-kube-api-access-xx4ms\") pod \"metallb-operator-controller-manager-f8cdcf667-bh7xq\" (UID: \"4e91721f-0e20-4dfb-9e17-45107d5cf8ce\") " pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.249087 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e91721f-0e20-4dfb-9e17-45107d5cf8ce-webhook-cert\") pod \"metallb-operator-controller-manager-f8cdcf667-bh7xq\" (UID: \"4e91721f-0e20-4dfb-9e17-45107d5cf8ce\") " pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.249122 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e91721f-0e20-4dfb-9e17-45107d5cf8ce-apiservice-cert\") pod \"metallb-operator-controller-manager-f8cdcf667-bh7xq\" (UID: \"4e91721f-0e20-4dfb-9e17-45107d5cf8ce\") " pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.350311 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e91721f-0e20-4dfb-9e17-45107d5cf8ce-webhook-cert\") pod \"metallb-operator-controller-manager-f8cdcf667-bh7xq\" (UID: \"4e91721f-0e20-4dfb-9e17-45107d5cf8ce\") " pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.350373 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e91721f-0e20-4dfb-9e17-45107d5cf8ce-apiservice-cert\") pod \"metallb-operator-controller-manager-f8cdcf667-bh7xq\" (UID: \"4e91721f-0e20-4dfb-9e17-45107d5cf8ce\") " pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.350451 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx4ms\" (UniqueName: \"kubernetes.io/projected/4e91721f-0e20-4dfb-9e17-45107d5cf8ce-kube-api-access-xx4ms\") pod \"metallb-operator-controller-manager-f8cdcf667-bh7xq\" (UID: \"4e91721f-0e20-4dfb-9e17-45107d5cf8ce\") " pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.366435 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e91721f-0e20-4dfb-9e17-45107d5cf8ce-webhook-cert\") pod \"metallb-operator-controller-manager-f8cdcf667-bh7xq\" (UID: \"4e91721f-0e20-4dfb-9e17-45107d5cf8ce\") " pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.373636 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e91721f-0e20-4dfb-9e17-45107d5cf8ce-apiservice-cert\") pod \"metallb-operator-controller-manager-f8cdcf667-bh7xq\" (UID: \"4e91721f-0e20-4dfb-9e17-45107d5cf8ce\") " pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.380698 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx4ms\" (UniqueName: \"kubernetes.io/projected/4e91721f-0e20-4dfb-9e17-45107d5cf8ce-kube-api-access-xx4ms\") pod \"metallb-operator-controller-manager-f8cdcf667-bh7xq\" (UID: \"4e91721f-0e20-4dfb-9e17-45107d5cf8ce\") " pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.475625 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.545718 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7"] Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.546576 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.549173 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.549714 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.551699 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-8z9f9" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.573522 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7"] Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.754703 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8787b83e-ce8e-4819-84f2-08adb1f238c8-webhook-cert\") pod \"metallb-operator-webhook-server-5c865dc449-g7xl7\" (UID: \"8787b83e-ce8e-4819-84f2-08adb1f238c8\") " pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.754835 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7vr7\" (UniqueName: \"kubernetes.io/projected/8787b83e-ce8e-4819-84f2-08adb1f238c8-kube-api-access-x7vr7\") pod \"metallb-operator-webhook-server-5c865dc449-g7xl7\" (UID: \"8787b83e-ce8e-4819-84f2-08adb1f238c8\") " pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.754946 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8787b83e-ce8e-4819-84f2-08adb1f238c8-apiservice-cert\") pod \"metallb-operator-webhook-server-5c865dc449-g7xl7\" (UID: \"8787b83e-ce8e-4819-84f2-08adb1f238c8\") " pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.856595 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8787b83e-ce8e-4819-84f2-08adb1f238c8-webhook-cert\") pod \"metallb-operator-webhook-server-5c865dc449-g7xl7\" (UID: \"8787b83e-ce8e-4819-84f2-08adb1f238c8\") " pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.857097 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7vr7\" (UniqueName: \"kubernetes.io/projected/8787b83e-ce8e-4819-84f2-08adb1f238c8-kube-api-access-x7vr7\") pod \"metallb-operator-webhook-server-5c865dc449-g7xl7\" (UID: \"8787b83e-ce8e-4819-84f2-08adb1f238c8\") " pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.857149 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8787b83e-ce8e-4819-84f2-08adb1f238c8-apiservice-cert\") pod \"metallb-operator-webhook-server-5c865dc449-g7xl7\" (UID: \"8787b83e-ce8e-4819-84f2-08adb1f238c8\") " pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.865933 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8787b83e-ce8e-4819-84f2-08adb1f238c8-webhook-cert\") pod \"metallb-operator-webhook-server-5c865dc449-g7xl7\" (UID: \"8787b83e-ce8e-4819-84f2-08adb1f238c8\") " pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.871466 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8787b83e-ce8e-4819-84f2-08adb1f238c8-apiservice-cert\") pod \"metallb-operator-webhook-server-5c865dc449-g7xl7\" (UID: \"8787b83e-ce8e-4819-84f2-08adb1f238c8\") " pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:35:44 crc kubenswrapper[4968]: I0218 15:35:44.875876 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7vr7\" (UniqueName: \"kubernetes.io/projected/8787b83e-ce8e-4819-84f2-08adb1f238c8-kube-api-access-x7vr7\") pod \"metallb-operator-webhook-server-5c865dc449-g7xl7\" (UID: \"8787b83e-ce8e-4819-84f2-08adb1f238c8\") " pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:35:45 crc kubenswrapper[4968]: I0218 15:35:45.066703 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq"] Feb 18 15:35:45 crc kubenswrapper[4968]: I0218 15:35:45.165116 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:35:45 crc kubenswrapper[4968]: I0218 15:35:45.660414 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7"] Feb 18 15:35:45 crc kubenswrapper[4968]: W0218 15:35:45.666144 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8787b83e_ce8e_4819_84f2_08adb1f238c8.slice/crio-61f31ce2435ed5a398f08d3c0c2a100b08aa6308cba0cc57957103a1eb6c6a4b WatchSource:0}: Error finding container 61f31ce2435ed5a398f08d3c0c2a100b08aa6308cba0cc57957103a1eb6c6a4b: Status 404 returned error can't find the container with id 61f31ce2435ed5a398f08d3c0c2a100b08aa6308cba0cc57957103a1eb6c6a4b Feb 18 15:35:45 crc kubenswrapper[4968]: I0218 15:35:45.711174 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" event={"ID":"4e91721f-0e20-4dfb-9e17-45107d5cf8ce","Type":"ContainerStarted","Data":"f5db01c4b32651afb9452af0a91024ee5a86fd25232283bd5d84c0f1d584a5c8"} Feb 18 15:35:45 crc kubenswrapper[4968]: I0218 15:35:45.712365 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" event={"ID":"8787b83e-ce8e-4819-84f2-08adb1f238c8","Type":"ContainerStarted","Data":"61f31ce2435ed5a398f08d3c0c2a100b08aa6308cba0cc57957103a1eb6c6a4b"} Feb 18 15:35:48 crc kubenswrapper[4968]: I0218 15:35:48.735911 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" event={"ID":"4e91721f-0e20-4dfb-9e17-45107d5cf8ce","Type":"ContainerStarted","Data":"610dce827aadd0663c71488f60e44725e9669521e8b452c474b6302dd4ed8c62"} Feb 18 15:35:48 crc kubenswrapper[4968]: I0218 15:35:48.737092 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:35:48 crc kubenswrapper[4968]: I0218 15:35:48.757624 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" podStartSLOduration=1.8140532569999999 podStartE2EDuration="4.757608967s" podCreationTimestamp="2026-02-18 15:35:44 +0000 UTC" firstStartedPulling="2026-02-18 15:35:45.058454339 +0000 UTC m=+844.443899211" lastFinishedPulling="2026-02-18 15:35:48.002010059 +0000 UTC m=+847.387454921" observedRunningTime="2026-02-18 15:35:48.755437116 +0000 UTC m=+848.140881978" watchObservedRunningTime="2026-02-18 15:35:48.757608967 +0000 UTC m=+848.143053829" Feb 18 15:35:50 crc kubenswrapper[4968]: I0218 15:35:50.749942 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" event={"ID":"8787b83e-ce8e-4819-84f2-08adb1f238c8","Type":"ContainerStarted","Data":"9a313ab9f9c933f0f57d3d1f5f6aee15f0bf54c105da02773249f3c030c96351"} Feb 18 15:35:50 crc kubenswrapper[4968]: I0218 15:35:50.750272 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:35:50 crc kubenswrapper[4968]: I0218 15:35:50.771059 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" podStartSLOduration=2.504582497 podStartE2EDuration="6.771039255s" podCreationTimestamp="2026-02-18 15:35:44 +0000 UTC" firstStartedPulling="2026-02-18 15:35:45.667767236 +0000 UTC m=+845.053212108" lastFinishedPulling="2026-02-18 15:35:49.934224004 +0000 UTC m=+849.319668866" observedRunningTime="2026-02-18 15:35:50.768067152 +0000 UTC m=+850.153512054" watchObservedRunningTime="2026-02-18 15:35:50.771039255 +0000 UTC m=+850.156484137" Feb 18 15:36:05 crc kubenswrapper[4968]: I0218 15:36:05.177251 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5c865dc449-g7xl7" Feb 18 15:36:24 crc kubenswrapper[4968]: I0218 15:36:24.480209 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-f8cdcf667-bh7xq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.208983 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-nz96q"] Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.212210 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.218469 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.219236 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z"] Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.219599 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.220454 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.222011 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.226731 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-8bhtv" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.243706 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z"] Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.268769 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-metrics-certs\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.268839 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jtdx\" (UniqueName: \"kubernetes.io/projected/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-kube-api-access-6jtdx\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.268880 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-metrics\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.268908 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-frr-conf\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.268930 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pmv7\" (UniqueName: \"kubernetes.io/projected/19821749-39d5-4ca0-8fb0-3b7edca2da5a-kube-api-access-9pmv7\") pod \"frr-k8s-webhook-server-78b44bf5bb-7c58z\" (UID: \"19821749-39d5-4ca0-8fb0-3b7edca2da5a\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.268966 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19821749-39d5-4ca0-8fb0-3b7edca2da5a-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7c58z\" (UID: \"19821749-39d5-4ca0-8fb0-3b7edca2da5a\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.268993 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-frr-startup\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.269030 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-reloader\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.269068 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-frr-sockets\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.306249 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-z4qpq"] Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.307205 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.321406 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.321811 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.321884 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.322085 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-vldgt" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.326996 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-69bbfbf88f-7ghrx"] Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.327924 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.329643 4968 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.331827 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-7ghrx"] Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.370696 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ab9fc6e-f33c-4351-b19a-638d53946dd0-cert\") pod \"controller-69bbfbf88f-7ghrx\" (UID: \"7ab9fc6e-f33c-4351-b19a-638d53946dd0\") " pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.370767 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgds4\" (UniqueName: \"kubernetes.io/projected/d3264f33-404d-4edb-a17a-1273abc843b5-kube-api-access-fgds4\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.370802 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-metrics\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.370830 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-metrics-certs\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.370862 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-frr-conf\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.370885 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pmv7\" (UniqueName: \"kubernetes.io/projected/19821749-39d5-4ca0-8fb0-3b7edca2da5a-kube-api-access-9pmv7\") pod \"frr-k8s-webhook-server-78b44bf5bb-7c58z\" (UID: \"19821749-39d5-4ca0-8fb0-3b7edca2da5a\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.370917 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19821749-39d5-4ca0-8fb0-3b7edca2da5a-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7c58z\" (UID: \"19821749-39d5-4ca0-8fb0-3b7edca2da5a\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.370945 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-frr-startup\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.370974 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdc82\" (UniqueName: \"kubernetes.io/projected/7ab9fc6e-f33c-4351-b19a-638d53946dd0-kube-api-access-bdc82\") pod \"controller-69bbfbf88f-7ghrx\" (UID: \"7ab9fc6e-f33c-4351-b19a-638d53946dd0\") " pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.371009 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-reloader\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.371035 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-memberlist\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.371070 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-frr-sockets\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.371101 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-metrics-certs\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.371133 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ab9fc6e-f33c-4351-b19a-638d53946dd0-metrics-certs\") pod \"controller-69bbfbf88f-7ghrx\" (UID: \"7ab9fc6e-f33c-4351-b19a-638d53946dd0\") " pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.371158 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d3264f33-404d-4edb-a17a-1273abc843b5-metallb-excludel2\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.371188 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jtdx\" (UniqueName: \"kubernetes.io/projected/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-kube-api-access-6jtdx\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.371317 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-metrics\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: E0218 15:36:25.371639 4968 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Feb 18 15:36:25 crc kubenswrapper[4968]: E0218 15:36:25.371691 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19821749-39d5-4ca0-8fb0-3b7edca2da5a-cert podName:19821749-39d5-4ca0-8fb0-3b7edca2da5a nodeName:}" failed. No retries permitted until 2026-02-18 15:36:25.87167546 +0000 UTC m=+885.257120322 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19821749-39d5-4ca0-8fb0-3b7edca2da5a-cert") pod "frr-k8s-webhook-server-78b44bf5bb-7c58z" (UID: "19821749-39d5-4ca0-8fb0-3b7edca2da5a") : secret "frr-k8s-webhook-server-cert" not found Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.371724 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-reloader\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.372004 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-frr-conf\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.372457 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-frr-startup\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.372626 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-frr-sockets\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.390590 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pmv7\" (UniqueName: \"kubernetes.io/projected/19821749-39d5-4ca0-8fb0-3b7edca2da5a-kube-api-access-9pmv7\") pod \"frr-k8s-webhook-server-78b44bf5bb-7c58z\" (UID: \"19821749-39d5-4ca0-8fb0-3b7edca2da5a\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.391013 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jtdx\" (UniqueName: \"kubernetes.io/projected/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-kube-api-access-6jtdx\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.393818 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56af3b3d-b4ba-4fe9-832f-969d3a767ed2-metrics-certs\") pod \"frr-k8s-nz96q\" (UID: \"56af3b3d-b4ba-4fe9-832f-969d3a767ed2\") " pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: E0218 15:36:25.472087 4968 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Feb 18 15:36:25 crc kubenswrapper[4968]: E0218 15:36:25.472156 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-metrics-certs podName:d3264f33-404d-4edb-a17a-1273abc843b5 nodeName:}" failed. No retries permitted until 2026-02-18 15:36:25.97213879 +0000 UTC m=+885.357583652 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-metrics-certs") pod "speaker-z4qpq" (UID: "d3264f33-404d-4edb-a17a-1273abc843b5") : secret "speaker-certs-secret" not found Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.472337 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-metrics-certs\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.472360 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgds4\" (UniqueName: \"kubernetes.io/projected/d3264f33-404d-4edb-a17a-1273abc843b5-kube-api-access-fgds4\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.472414 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdc82\" (UniqueName: \"kubernetes.io/projected/7ab9fc6e-f33c-4351-b19a-638d53946dd0-kube-api-access-bdc82\") pod \"controller-69bbfbf88f-7ghrx\" (UID: \"7ab9fc6e-f33c-4351-b19a-638d53946dd0\") " pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.472441 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-memberlist\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.472479 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ab9fc6e-f33c-4351-b19a-638d53946dd0-metrics-certs\") pod \"controller-69bbfbf88f-7ghrx\" (UID: \"7ab9fc6e-f33c-4351-b19a-638d53946dd0\") " pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.472497 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d3264f33-404d-4edb-a17a-1273abc843b5-metallb-excludel2\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.472528 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ab9fc6e-f33c-4351-b19a-638d53946dd0-cert\") pod \"controller-69bbfbf88f-7ghrx\" (UID: \"7ab9fc6e-f33c-4351-b19a-638d53946dd0\") " pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:25 crc kubenswrapper[4968]: E0218 15:36:25.473096 4968 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 18 15:36:25 crc kubenswrapper[4968]: E0218 15:36:25.473180 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-memberlist podName:d3264f33-404d-4edb-a17a-1273abc843b5 nodeName:}" failed. No retries permitted until 2026-02-18 15:36:25.973155389 +0000 UTC m=+885.358600251 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-memberlist") pod "speaker-z4qpq" (UID: "d3264f33-404d-4edb-a17a-1273abc843b5") : secret "metallb-memberlist" not found Feb 18 15:36:25 crc kubenswrapper[4968]: E0218 15:36:25.473237 4968 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Feb 18 15:36:25 crc kubenswrapper[4968]: E0218 15:36:25.473322 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7ab9fc6e-f33c-4351-b19a-638d53946dd0-metrics-certs podName:7ab9fc6e-f33c-4351-b19a-638d53946dd0 nodeName:}" failed. No retries permitted until 2026-02-18 15:36:25.973299163 +0000 UTC m=+885.358744115 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7ab9fc6e-f33c-4351-b19a-638d53946dd0-metrics-certs") pod "controller-69bbfbf88f-7ghrx" (UID: "7ab9fc6e-f33c-4351-b19a-638d53946dd0") : secret "controller-certs-secret" not found Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.473564 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d3264f33-404d-4edb-a17a-1273abc843b5-metallb-excludel2\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.476946 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ab9fc6e-f33c-4351-b19a-638d53946dd0-cert\") pod \"controller-69bbfbf88f-7ghrx\" (UID: \"7ab9fc6e-f33c-4351-b19a-638d53946dd0\") " pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.487707 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdc82\" (UniqueName: \"kubernetes.io/projected/7ab9fc6e-f33c-4351-b19a-638d53946dd0-kube-api-access-bdc82\") pod \"controller-69bbfbf88f-7ghrx\" (UID: \"7ab9fc6e-f33c-4351-b19a-638d53946dd0\") " pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.490848 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgds4\" (UniqueName: \"kubernetes.io/projected/d3264f33-404d-4edb-a17a-1273abc843b5-kube-api-access-fgds4\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.533379 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.877594 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19821749-39d5-4ca0-8fb0-3b7edca2da5a-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7c58z\" (UID: \"19821749-39d5-4ca0-8fb0-3b7edca2da5a\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.883556 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19821749-39d5-4ca0-8fb0-3b7edca2da5a-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-7c58z\" (UID: \"19821749-39d5-4ca0-8fb0-3b7edca2da5a\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.979366 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ab9fc6e-f33c-4351-b19a-638d53946dd0-metrics-certs\") pod \"controller-69bbfbf88f-7ghrx\" (UID: \"7ab9fc6e-f33c-4351-b19a-638d53946dd0\") " pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.980367 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-metrics-certs\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.980685 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-memberlist\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: E0218 15:36:25.980956 4968 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 18 15:36:25 crc kubenswrapper[4968]: E0218 15:36:25.981052 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-memberlist podName:d3264f33-404d-4edb-a17a-1273abc843b5 nodeName:}" failed. No retries permitted until 2026-02-18 15:36:26.981021447 +0000 UTC m=+886.366466339 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-memberlist") pod "speaker-z4qpq" (UID: "d3264f33-404d-4edb-a17a-1273abc843b5") : secret "metallb-memberlist" not found Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.983069 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nz96q" event={"ID":"56af3b3d-b4ba-4fe9-832f-969d3a767ed2","Type":"ContainerStarted","Data":"93e64af7f1a4ecd65ec6e94c8999af274b0a2b18d309e9f47e0c8e5c27ce1dcf"} Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.984587 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-metrics-certs\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:25 crc kubenswrapper[4968]: I0218 15:36:25.985371 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ab9fc6e-f33c-4351-b19a-638d53946dd0-metrics-certs\") pod \"controller-69bbfbf88f-7ghrx\" (UID: \"7ab9fc6e-f33c-4351-b19a-638d53946dd0\") " pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:26 crc kubenswrapper[4968]: I0218 15:36:26.141463 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" Feb 18 15:36:26 crc kubenswrapper[4968]: I0218 15:36:26.241192 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:26 crc kubenswrapper[4968]: I0218 15:36:26.408342 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z"] Feb 18 15:36:26 crc kubenswrapper[4968]: I0218 15:36:26.651324 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-7ghrx"] Feb 18 15:36:26 crc kubenswrapper[4968]: W0218 15:36:26.658959 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ab9fc6e_f33c_4351_b19a_638d53946dd0.slice/crio-de2605807039080aa5009895c4e3030d53801e84b7195b846dfc0f3335ec1112 WatchSource:0}: Error finding container de2605807039080aa5009895c4e3030d53801e84b7195b846dfc0f3335ec1112: Status 404 returned error can't find the container with id de2605807039080aa5009895c4e3030d53801e84b7195b846dfc0f3335ec1112 Feb 18 15:36:26 crc kubenswrapper[4968]: I0218 15:36:26.992252 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-7ghrx" event={"ID":"7ab9fc6e-f33c-4351-b19a-638d53946dd0","Type":"ContainerStarted","Data":"06e0f699b3d088dec3db26e9c432c58442e11d8145dfc816631f7c8e4a9c4476"} Feb 18 15:36:26 crc kubenswrapper[4968]: I0218 15:36:26.992497 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-7ghrx" event={"ID":"7ab9fc6e-f33c-4351-b19a-638d53946dd0","Type":"ContainerStarted","Data":"de2605807039080aa5009895c4e3030d53801e84b7195b846dfc0f3335ec1112"} Feb 18 15:36:26 crc kubenswrapper[4968]: I0218 15:36:26.993572 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" event={"ID":"19821749-39d5-4ca0-8fb0-3b7edca2da5a","Type":"ContainerStarted","Data":"ab5fa1fd68d0218e2ce7b1f420e02f1877c79100317c26d6de1cdfdd3ab886ce"} Feb 18 15:36:26 crc kubenswrapper[4968]: I0218 15:36:26.993905 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-memberlist\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:26 crc kubenswrapper[4968]: I0218 15:36:26.998914 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3264f33-404d-4edb-a17a-1273abc843b5-memberlist\") pod \"speaker-z4qpq\" (UID: \"d3264f33-404d-4edb-a17a-1273abc843b5\") " pod="metallb-system/speaker-z4qpq" Feb 18 15:36:27 crc kubenswrapper[4968]: I0218 15:36:27.121465 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-z4qpq" Feb 18 15:36:27 crc kubenswrapper[4968]: W0218 15:36:27.141477 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3264f33_404d_4edb_a17a_1273abc843b5.slice/crio-a2c2916f897db8084f44efd7d85743725b6aadfc227135eb2b520551ee44f6c1 WatchSource:0}: Error finding container a2c2916f897db8084f44efd7d85743725b6aadfc227135eb2b520551ee44f6c1: Status 404 returned error can't find the container with id a2c2916f897db8084f44efd7d85743725b6aadfc227135eb2b520551ee44f6c1 Feb 18 15:36:28 crc kubenswrapper[4968]: I0218 15:36:28.009489 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-7ghrx" event={"ID":"7ab9fc6e-f33c-4351-b19a-638d53946dd0","Type":"ContainerStarted","Data":"d6cfe70f093f2104e731c7c125f163a49de14cbafdb29e5124a5104c19a9356a"} Feb 18 15:36:28 crc kubenswrapper[4968]: I0218 15:36:28.009810 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:28 crc kubenswrapper[4968]: I0218 15:36:28.015844 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-z4qpq" event={"ID":"d3264f33-404d-4edb-a17a-1273abc843b5","Type":"ContainerStarted","Data":"4dd05a014bb58900bba54f9c0011c1fcf27c06c29e9f06183904551a5c704876"} Feb 18 15:36:28 crc kubenswrapper[4968]: I0218 15:36:28.015893 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-z4qpq" event={"ID":"d3264f33-404d-4edb-a17a-1273abc843b5","Type":"ContainerStarted","Data":"1508bea8f28ea9c764ad605d54a8f84210395305900a2dd2f9e2c4806fa5a572"} Feb 18 15:36:28 crc kubenswrapper[4968]: I0218 15:36:28.015904 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-z4qpq" event={"ID":"d3264f33-404d-4edb-a17a-1273abc843b5","Type":"ContainerStarted","Data":"a2c2916f897db8084f44efd7d85743725b6aadfc227135eb2b520551ee44f6c1"} Feb 18 15:36:28 crc kubenswrapper[4968]: I0218 15:36:28.016090 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-z4qpq" Feb 18 15:36:28 crc kubenswrapper[4968]: I0218 15:36:28.042107 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-69bbfbf88f-7ghrx" podStartSLOduration=3.042086057 podStartE2EDuration="3.042086057s" podCreationTimestamp="2026-02-18 15:36:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:36:28.036604834 +0000 UTC m=+887.422049696" watchObservedRunningTime="2026-02-18 15:36:28.042086057 +0000 UTC m=+887.427530919" Feb 18 15:36:28 crc kubenswrapper[4968]: I0218 15:36:28.061405 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-z4qpq" podStartSLOduration=3.061387487 podStartE2EDuration="3.061387487s" podCreationTimestamp="2026-02-18 15:36:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:36:28.057989762 +0000 UTC m=+887.443434624" watchObservedRunningTime="2026-02-18 15:36:28.061387487 +0000 UTC m=+887.446832349" Feb 18 15:36:33 crc kubenswrapper[4968]: I0218 15:36:33.067003 4968 generic.go:334] "Generic (PLEG): container finished" podID="56af3b3d-b4ba-4fe9-832f-969d3a767ed2" containerID="553873d95361dce7511b2b0039b89ff498a67b1fb8a8a25e7d40df7813cde027" exitCode=0 Feb 18 15:36:33 crc kubenswrapper[4968]: I0218 15:36:33.067058 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nz96q" event={"ID":"56af3b3d-b4ba-4fe9-832f-969d3a767ed2","Type":"ContainerDied","Data":"553873d95361dce7511b2b0039b89ff498a67b1fb8a8a25e7d40df7813cde027"} Feb 18 15:36:33 crc kubenswrapper[4968]: I0218 15:36:33.085807 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" event={"ID":"19821749-39d5-4ca0-8fb0-3b7edca2da5a","Type":"ContainerStarted","Data":"256707e985ec29856eebb9bdf2a360db487c2f1a6f66f95db74235199df82976"} Feb 18 15:36:33 crc kubenswrapper[4968]: I0218 15:36:33.085998 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" Feb 18 15:36:33 crc kubenswrapper[4968]: I0218 15:36:33.148589 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" podStartSLOduration=1.764831675 podStartE2EDuration="8.148562816s" podCreationTimestamp="2026-02-18 15:36:25 +0000 UTC" firstStartedPulling="2026-02-18 15:36:26.415915214 +0000 UTC m=+885.801360076" lastFinishedPulling="2026-02-18 15:36:32.799646355 +0000 UTC m=+892.185091217" observedRunningTime="2026-02-18 15:36:33.144019229 +0000 UTC m=+892.529464101" watchObservedRunningTime="2026-02-18 15:36:33.148562816 +0000 UTC m=+892.534007718" Feb 18 15:36:34 crc kubenswrapper[4968]: I0218 15:36:34.095712 4968 generic.go:334] "Generic (PLEG): container finished" podID="56af3b3d-b4ba-4fe9-832f-969d3a767ed2" containerID="fda2949e6a24c06ef3f31ab04d85b72573e73a8b058e35b5a43b391b04374907" exitCode=0 Feb 18 15:36:34 crc kubenswrapper[4968]: I0218 15:36:34.096364 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nz96q" event={"ID":"56af3b3d-b4ba-4fe9-832f-969d3a767ed2","Type":"ContainerDied","Data":"fda2949e6a24c06ef3f31ab04d85b72573e73a8b058e35b5a43b391b04374907"} Feb 18 15:36:35 crc kubenswrapper[4968]: I0218 15:36:35.108329 4968 generic.go:334] "Generic (PLEG): container finished" podID="56af3b3d-b4ba-4fe9-832f-969d3a767ed2" containerID="c8fc1c5abcc13f43a713d87487c78d31ed4e9c76596f4e9e27ba4c0d5f1d42cd" exitCode=0 Feb 18 15:36:35 crc kubenswrapper[4968]: I0218 15:36:35.108663 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nz96q" event={"ID":"56af3b3d-b4ba-4fe9-832f-969d3a767ed2","Type":"ContainerDied","Data":"c8fc1c5abcc13f43a713d87487c78d31ed4e9c76596f4e9e27ba4c0d5f1d42cd"} Feb 18 15:36:36 crc kubenswrapper[4968]: I0218 15:36:36.123701 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nz96q" event={"ID":"56af3b3d-b4ba-4fe9-832f-969d3a767ed2","Type":"ContainerStarted","Data":"a212fa3cea0d52c52b52ac8da005796141a2ecee87e0aa103475a52ab40f5895"} Feb 18 15:36:36 crc kubenswrapper[4968]: I0218 15:36:36.124164 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nz96q" event={"ID":"56af3b3d-b4ba-4fe9-832f-969d3a767ed2","Type":"ContainerStarted","Data":"e5bc9a00dbee179fcf72033a6df568315cd23f1518bcc71bd55ca620b84308d7"} Feb 18 15:36:36 crc kubenswrapper[4968]: I0218 15:36:36.124184 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nz96q" event={"ID":"56af3b3d-b4ba-4fe9-832f-969d3a767ed2","Type":"ContainerStarted","Data":"c9278d84459f3e16999bad82454b6b171b39fdd50579d21d74c19ba12dd49fa6"} Feb 18 15:36:36 crc kubenswrapper[4968]: I0218 15:36:36.124200 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nz96q" event={"ID":"56af3b3d-b4ba-4fe9-832f-969d3a767ed2","Type":"ContainerStarted","Data":"a73e3601466ef5d31a197b1d7aabd8e6fac54a3ebe292b5fc8c1fe9e0422f46d"} Feb 18 15:36:36 crc kubenswrapper[4968]: I0218 15:36:36.124216 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nz96q" event={"ID":"56af3b3d-b4ba-4fe9-832f-969d3a767ed2","Type":"ContainerStarted","Data":"23aff99e5cc2732f6eabebf78b3cc0972068828d995e9dd6b6f8e13fa63b7136"} Feb 18 15:36:36 crc kubenswrapper[4968]: I0218 15:36:36.245602 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-69bbfbf88f-7ghrx" Feb 18 15:36:37 crc kubenswrapper[4968]: I0218 15:36:37.128519 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-z4qpq" Feb 18 15:36:37 crc kubenswrapper[4968]: I0218 15:36:37.135373 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-nz96q" event={"ID":"56af3b3d-b4ba-4fe9-832f-969d3a767ed2","Type":"ContainerStarted","Data":"5e12789e084b5fc3e443bf9d2d98669dc78ad96772b79a85a471934f4bef6cba"} Feb 18 15:36:37 crc kubenswrapper[4968]: I0218 15:36:37.135564 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:37 crc kubenswrapper[4968]: I0218 15:36:37.204912 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-nz96q" podStartSLOduration=5.070875064 podStartE2EDuration="12.204888775s" podCreationTimestamp="2026-02-18 15:36:25 +0000 UTC" firstStartedPulling="2026-02-18 15:36:25.65190643 +0000 UTC m=+885.037351302" lastFinishedPulling="2026-02-18 15:36:32.785920121 +0000 UTC m=+892.171365013" observedRunningTime="2026-02-18 15:36:37.202463978 +0000 UTC m=+896.587908840" watchObservedRunningTime="2026-02-18 15:36:37.204888775 +0000 UTC m=+896.590333637" Feb 18 15:36:39 crc kubenswrapper[4968]: I0218 15:36:39.790260 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-4dqr6"] Feb 18 15:36:39 crc kubenswrapper[4968]: I0218 15:36:39.791444 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4dqr6" Feb 18 15:36:39 crc kubenswrapper[4968]: I0218 15:36:39.799023 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-fk5ph" Feb 18 15:36:39 crc kubenswrapper[4968]: I0218 15:36:39.799226 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 18 15:36:39 crc kubenswrapper[4968]: I0218 15:36:39.806334 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 18 15:36:39 crc kubenswrapper[4968]: I0218 15:36:39.812055 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4dqr6"] Feb 18 15:36:39 crc kubenswrapper[4968]: I0218 15:36:39.877741 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvv5n\" (UniqueName: \"kubernetes.io/projected/753a7804-4ee1-44d1-8ed1-e66d0eac1897-kube-api-access-pvv5n\") pod \"openstack-operator-index-4dqr6\" (UID: \"753a7804-4ee1-44d1-8ed1-e66d0eac1897\") " pod="openstack-operators/openstack-operator-index-4dqr6" Feb 18 15:36:39 crc kubenswrapper[4968]: I0218 15:36:39.979687 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvv5n\" (UniqueName: \"kubernetes.io/projected/753a7804-4ee1-44d1-8ed1-e66d0eac1897-kube-api-access-pvv5n\") pod \"openstack-operator-index-4dqr6\" (UID: \"753a7804-4ee1-44d1-8ed1-e66d0eac1897\") " pod="openstack-operators/openstack-operator-index-4dqr6" Feb 18 15:36:39 crc kubenswrapper[4968]: I0218 15:36:39.998240 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvv5n\" (UniqueName: \"kubernetes.io/projected/753a7804-4ee1-44d1-8ed1-e66d0eac1897-kube-api-access-pvv5n\") pod \"openstack-operator-index-4dqr6\" (UID: \"753a7804-4ee1-44d1-8ed1-e66d0eac1897\") " pod="openstack-operators/openstack-operator-index-4dqr6" Feb 18 15:36:40 crc kubenswrapper[4968]: I0218 15:36:40.112899 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4dqr6" Feb 18 15:36:40 crc kubenswrapper[4968]: I0218 15:36:40.313008 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4dqr6"] Feb 18 15:36:40 crc kubenswrapper[4968]: W0218 15:36:40.320029 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod753a7804_4ee1_44d1_8ed1_e66d0eac1897.slice/crio-da9dcbbf0db52ea89ce3cec6cb952bc3a8cbe2505a52eec80f27548fe9649e6f WatchSource:0}: Error finding container da9dcbbf0db52ea89ce3cec6cb952bc3a8cbe2505a52eec80f27548fe9649e6f: Status 404 returned error can't find the container with id da9dcbbf0db52ea89ce3cec6cb952bc3a8cbe2505a52eec80f27548fe9649e6f Feb 18 15:36:40 crc kubenswrapper[4968]: I0218 15:36:40.534311 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:40 crc kubenswrapper[4968]: I0218 15:36:40.595252 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:41 crc kubenswrapper[4968]: I0218 15:36:41.162371 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4dqr6" event={"ID":"753a7804-4ee1-44d1-8ed1-e66d0eac1897","Type":"ContainerStarted","Data":"da9dcbbf0db52ea89ce3cec6cb952bc3a8cbe2505a52eec80f27548fe9649e6f"} Feb 18 15:36:43 crc kubenswrapper[4968]: I0218 15:36:43.766680 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-4dqr6"] Feb 18 15:36:44 crc kubenswrapper[4968]: I0218 15:36:44.189690 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4dqr6" event={"ID":"753a7804-4ee1-44d1-8ed1-e66d0eac1897","Type":"ContainerStarted","Data":"28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002"} Feb 18 15:36:44 crc kubenswrapper[4968]: I0218 15:36:44.208319 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-4dqr6" podStartSLOduration=2.048451693 podStartE2EDuration="5.208293632s" podCreationTimestamp="2026-02-18 15:36:39 +0000 UTC" firstStartedPulling="2026-02-18 15:36:40.321549527 +0000 UTC m=+899.706994399" lastFinishedPulling="2026-02-18 15:36:43.481391476 +0000 UTC m=+902.866836338" observedRunningTime="2026-02-18 15:36:44.207405907 +0000 UTC m=+903.592850799" watchObservedRunningTime="2026-02-18 15:36:44.208293632 +0000 UTC m=+903.593738514" Feb 18 15:36:44 crc kubenswrapper[4968]: I0218 15:36:44.367795 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:36:44 crc kubenswrapper[4968]: I0218 15:36:44.367881 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:36:44 crc kubenswrapper[4968]: I0218 15:36:44.581166 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-b6km5"] Feb 18 15:36:44 crc kubenswrapper[4968]: I0218 15:36:44.582947 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-b6km5" Feb 18 15:36:44 crc kubenswrapper[4968]: I0218 15:36:44.593682 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-b6km5"] Feb 18 15:36:44 crc kubenswrapper[4968]: I0218 15:36:44.648431 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27267\" (UniqueName: \"kubernetes.io/projected/858314b2-7c36-467c-a1a5-dd764bd22784-kube-api-access-27267\") pod \"openstack-operator-index-b6km5\" (UID: \"858314b2-7c36-467c-a1a5-dd764bd22784\") " pod="openstack-operators/openstack-operator-index-b6km5" Feb 18 15:36:44 crc kubenswrapper[4968]: I0218 15:36:44.749616 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27267\" (UniqueName: \"kubernetes.io/projected/858314b2-7c36-467c-a1a5-dd764bd22784-kube-api-access-27267\") pod \"openstack-operator-index-b6km5\" (UID: \"858314b2-7c36-467c-a1a5-dd764bd22784\") " pod="openstack-operators/openstack-operator-index-b6km5" Feb 18 15:36:44 crc kubenswrapper[4968]: I0218 15:36:44.780475 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27267\" (UniqueName: \"kubernetes.io/projected/858314b2-7c36-467c-a1a5-dd764bd22784-kube-api-access-27267\") pod \"openstack-operator-index-b6km5\" (UID: \"858314b2-7c36-467c-a1a5-dd764bd22784\") " pod="openstack-operators/openstack-operator-index-b6km5" Feb 18 15:36:44 crc kubenswrapper[4968]: I0218 15:36:44.903214 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-b6km5" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.194914 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-4dqr6" podUID="753a7804-4ee1-44d1-8ed1-e66d0eac1897" containerName="registry-server" containerID="cri-o://28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002" gracePeriod=2 Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.421905 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-b6km5"] Feb 18 15:36:45 crc kubenswrapper[4968]: W0218 15:36:45.436559 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod858314b2_7c36_467c_a1a5_dd764bd22784.slice/crio-44940668ff2df532eba1d3564964a2b4f0e81e39606396ebab52418e554d1a54 WatchSource:0}: Error finding container 44940668ff2df532eba1d3564964a2b4f0e81e39606396ebab52418e554d1a54: Status 404 returned error can't find the container with id 44940668ff2df532eba1d3564964a2b4f0e81e39606396ebab52418e554d1a54 Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.538141 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-nz96q" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.597181 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jh62p"] Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.599440 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.599764 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jh62p"] Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.639865 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4dqr6" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.673385 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvv5n\" (UniqueName: \"kubernetes.io/projected/753a7804-4ee1-44d1-8ed1-e66d0eac1897-kube-api-access-pvv5n\") pod \"753a7804-4ee1-44d1-8ed1-e66d0eac1897\" (UID: \"753a7804-4ee1-44d1-8ed1-e66d0eac1897\") " Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.673601 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-utilities\") pod \"community-operators-jh62p\" (UID: \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\") " pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.673634 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-catalog-content\") pod \"community-operators-jh62p\" (UID: \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\") " pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.673680 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdcpc\" (UniqueName: \"kubernetes.io/projected/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-kube-api-access-cdcpc\") pod \"community-operators-jh62p\" (UID: \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\") " pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.679065 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/753a7804-4ee1-44d1-8ed1-e66d0eac1897-kube-api-access-pvv5n" (OuterVolumeSpecName: "kube-api-access-pvv5n") pod "753a7804-4ee1-44d1-8ed1-e66d0eac1897" (UID: "753a7804-4ee1-44d1-8ed1-e66d0eac1897"). InnerVolumeSpecName "kube-api-access-pvv5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.774870 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-utilities\") pod \"community-operators-jh62p\" (UID: \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\") " pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.774956 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-catalog-content\") pod \"community-operators-jh62p\" (UID: \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\") " pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.775042 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdcpc\" (UniqueName: \"kubernetes.io/projected/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-kube-api-access-cdcpc\") pod \"community-operators-jh62p\" (UID: \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\") " pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.775166 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvv5n\" (UniqueName: \"kubernetes.io/projected/753a7804-4ee1-44d1-8ed1-e66d0eac1897-kube-api-access-pvv5n\") on node \"crc\" DevicePath \"\"" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.776200 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-utilities\") pod \"community-operators-jh62p\" (UID: \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\") " pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.776573 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-catalog-content\") pod \"community-operators-jh62p\" (UID: \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\") " pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.791473 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdcpc\" (UniqueName: \"kubernetes.io/projected/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-kube-api-access-cdcpc\") pod \"community-operators-jh62p\" (UID: \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\") " pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:45 crc kubenswrapper[4968]: I0218 15:36:45.929327 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.155112 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-7c58z" Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.205907 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-b6km5" event={"ID":"858314b2-7c36-467c-a1a5-dd764bd22784","Type":"ContainerStarted","Data":"743dc8fd1464b7b1ba46579ab14b754a1515b7e9ee05680a6a751645446f1d23"} Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.205953 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-b6km5" event={"ID":"858314b2-7c36-467c-a1a5-dd764bd22784","Type":"ContainerStarted","Data":"44940668ff2df532eba1d3564964a2b4f0e81e39606396ebab52418e554d1a54"} Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.208016 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jh62p"] Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.208527 4968 generic.go:334] "Generic (PLEG): container finished" podID="753a7804-4ee1-44d1-8ed1-e66d0eac1897" containerID="28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002" exitCode=0 Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.208552 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4dqr6" event={"ID":"753a7804-4ee1-44d1-8ed1-e66d0eac1897","Type":"ContainerDied","Data":"28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002"} Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.208571 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4dqr6" event={"ID":"753a7804-4ee1-44d1-8ed1-e66d0eac1897","Type":"ContainerDied","Data":"da9dcbbf0db52ea89ce3cec6cb952bc3a8cbe2505a52eec80f27548fe9649e6f"} Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.208587 4968 scope.go:117] "RemoveContainer" containerID="28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002" Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.208716 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4dqr6" Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.226363 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-b6km5" podStartSLOduration=2.162436311 podStartE2EDuration="2.226346619s" podCreationTimestamp="2026-02-18 15:36:44 +0000 UTC" firstStartedPulling="2026-02-18 15:36:45.440925177 +0000 UTC m=+904.826370079" lastFinishedPulling="2026-02-18 15:36:45.504835525 +0000 UTC m=+904.890280387" observedRunningTime="2026-02-18 15:36:46.226003089 +0000 UTC m=+905.611447951" watchObservedRunningTime="2026-02-18 15:36:46.226346619 +0000 UTC m=+905.611791481" Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.262795 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-4dqr6"] Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.265772 4968 scope.go:117] "RemoveContainer" containerID="28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002" Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.265867 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-4dqr6"] Feb 18 15:36:46 crc kubenswrapper[4968]: E0218 15:36:46.266283 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002\": container with ID starting with 28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002 not found: ID does not exist" containerID="28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002" Feb 18 15:36:46 crc kubenswrapper[4968]: I0218 15:36:46.266311 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002"} err="failed to get container status \"28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002\": rpc error: code = NotFound desc = could not find container \"28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002\": container with ID starting with 28c9f27452c87ac6c4fa5e7f11c85ec6991813cb3ba180b90d7e9333bbfa2002 not found: ID does not exist" Feb 18 15:36:47 crc kubenswrapper[4968]: I0218 15:36:47.222045 4968 generic.go:334] "Generic (PLEG): container finished" podID="735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" containerID="495db75894ccecb54599d599c2b306f19e348a253687494714687b6d51d387c3" exitCode=0 Feb 18 15:36:47 crc kubenswrapper[4968]: I0218 15:36:47.222128 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh62p" event={"ID":"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d","Type":"ContainerDied","Data":"495db75894ccecb54599d599c2b306f19e348a253687494714687b6d51d387c3"} Feb 18 15:36:47 crc kubenswrapper[4968]: I0218 15:36:47.222510 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh62p" event={"ID":"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d","Type":"ContainerStarted","Data":"7eb6f9a688fcde99fecd57f8cb3462b17202c1ba0d02aed858cd571c82e6d259"} Feb 18 15:36:47 crc kubenswrapper[4968]: I0218 15:36:47.252693 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="753a7804-4ee1-44d1-8ed1-e66d0eac1897" path="/var/lib/kubelet/pods/753a7804-4ee1-44d1-8ed1-e66d0eac1897/volumes" Feb 18 15:36:48 crc kubenswrapper[4968]: I0218 15:36:48.233074 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh62p" event={"ID":"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d","Type":"ContainerStarted","Data":"60990d97f8cabe1206fd797563fce683777c6feb43844093a225897610bebec1"} Feb 18 15:36:49 crc kubenswrapper[4968]: I0218 15:36:49.242966 4968 generic.go:334] "Generic (PLEG): container finished" podID="735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" containerID="60990d97f8cabe1206fd797563fce683777c6feb43844093a225897610bebec1" exitCode=0 Feb 18 15:36:49 crc kubenswrapper[4968]: I0218 15:36:49.243014 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh62p" event={"ID":"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d","Type":"ContainerDied","Data":"60990d97f8cabe1206fd797563fce683777c6feb43844093a225897610bebec1"} Feb 18 15:36:50 crc kubenswrapper[4968]: I0218 15:36:50.253124 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh62p" event={"ID":"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d","Type":"ContainerStarted","Data":"123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc"} Feb 18 15:36:50 crc kubenswrapper[4968]: I0218 15:36:50.277671 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jh62p" podStartSLOduration=2.910593167 podStartE2EDuration="5.277652148s" podCreationTimestamp="2026-02-18 15:36:45 +0000 UTC" firstStartedPulling="2026-02-18 15:36:47.228061213 +0000 UTC m=+906.613506115" lastFinishedPulling="2026-02-18 15:36:49.595120224 +0000 UTC m=+908.980565096" observedRunningTime="2026-02-18 15:36:50.277392991 +0000 UTC m=+909.662837863" watchObservedRunningTime="2026-02-18 15:36:50.277652148 +0000 UTC m=+909.663097020" Feb 18 15:36:54 crc kubenswrapper[4968]: I0218 15:36:54.904035 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-b6km5" Feb 18 15:36:54 crc kubenswrapper[4968]: I0218 15:36:54.904132 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-b6km5" Feb 18 15:36:54 crc kubenswrapper[4968]: I0218 15:36:54.946506 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-b6km5" Feb 18 15:36:55 crc kubenswrapper[4968]: I0218 15:36:55.327037 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-b6km5" Feb 18 15:36:55 crc kubenswrapper[4968]: I0218 15:36:55.930252 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:55 crc kubenswrapper[4968]: I0218 15:36:55.930369 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:56 crc kubenswrapper[4968]: I0218 15:36:56.004409 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:56 crc kubenswrapper[4968]: I0218 15:36:56.387581 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.201890 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2"] Feb 18 15:36:57 crc kubenswrapper[4968]: E0218 15:36:57.202247 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="753a7804-4ee1-44d1-8ed1-e66d0eac1897" containerName="registry-server" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.202264 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="753a7804-4ee1-44d1-8ed1-e66d0eac1897" containerName="registry-server" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.202441 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="753a7804-4ee1-44d1-8ed1-e66d0eac1897" containerName="registry-server" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.203601 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.207695 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-s72bd" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.212861 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2"] Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.247078 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a777cdac-6acd-45d2-9262-8c77577fe2b6-bundle\") pod \"74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2\" (UID: \"a777cdac-6acd-45d2-9262-8c77577fe2b6\") " pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.247129 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v845k\" (UniqueName: \"kubernetes.io/projected/a777cdac-6acd-45d2-9262-8c77577fe2b6-kube-api-access-v845k\") pod \"74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2\" (UID: \"a777cdac-6acd-45d2-9262-8c77577fe2b6\") " pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.247156 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a777cdac-6acd-45d2-9262-8c77577fe2b6-util\") pod \"74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2\" (UID: \"a777cdac-6acd-45d2-9262-8c77577fe2b6\") " pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.348142 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a777cdac-6acd-45d2-9262-8c77577fe2b6-bundle\") pod \"74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2\" (UID: \"a777cdac-6acd-45d2-9262-8c77577fe2b6\") " pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.348213 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v845k\" (UniqueName: \"kubernetes.io/projected/a777cdac-6acd-45d2-9262-8c77577fe2b6-kube-api-access-v845k\") pod \"74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2\" (UID: \"a777cdac-6acd-45d2-9262-8c77577fe2b6\") " pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.348243 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a777cdac-6acd-45d2-9262-8c77577fe2b6-util\") pod \"74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2\" (UID: \"a777cdac-6acd-45d2-9262-8c77577fe2b6\") " pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.349621 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a777cdac-6acd-45d2-9262-8c77577fe2b6-util\") pod \"74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2\" (UID: \"a777cdac-6acd-45d2-9262-8c77577fe2b6\") " pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.349910 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a777cdac-6acd-45d2-9262-8c77577fe2b6-bundle\") pod \"74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2\" (UID: \"a777cdac-6acd-45d2-9262-8c77577fe2b6\") " pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.373926 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v845k\" (UniqueName: \"kubernetes.io/projected/a777cdac-6acd-45d2-9262-8c77577fe2b6-kube-api-access-v845k\") pod \"74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2\" (UID: \"a777cdac-6acd-45d2-9262-8c77577fe2b6\") " pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.529596 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:36:57 crc kubenswrapper[4968]: I0218 15:36:57.962693 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jh62p"] Feb 18 15:36:58 crc kubenswrapper[4968]: I0218 15:36:58.053772 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2"] Feb 18 15:36:58 crc kubenswrapper[4968]: W0218 15:36:58.056582 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda777cdac_6acd_45d2_9262_8c77577fe2b6.slice/crio-3faa7d11d9f728fa945a3ddb4d3a8b5f62ae326470926833b26fae43cd853289 WatchSource:0}: Error finding container 3faa7d11d9f728fa945a3ddb4d3a8b5f62ae326470926833b26fae43cd853289: Status 404 returned error can't find the container with id 3faa7d11d9f728fa945a3ddb4d3a8b5f62ae326470926833b26fae43cd853289 Feb 18 15:36:58 crc kubenswrapper[4968]: I0218 15:36:58.319475 4968 generic.go:334] "Generic (PLEG): container finished" podID="a777cdac-6acd-45d2-9262-8c77577fe2b6" containerID="7c2947c708ba20da823ddd665b4f90ffea536c19f4f0562b0b34c89cc69fb6bf" exitCode=0 Feb 18 15:36:58 crc kubenswrapper[4968]: I0218 15:36:58.319612 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" event={"ID":"a777cdac-6acd-45d2-9262-8c77577fe2b6","Type":"ContainerDied","Data":"7c2947c708ba20da823ddd665b4f90ffea536c19f4f0562b0b34c89cc69fb6bf"} Feb 18 15:36:58 crc kubenswrapper[4968]: I0218 15:36:58.319682 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jh62p" podUID="735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" containerName="registry-server" containerID="cri-o://123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc" gracePeriod=2 Feb 18 15:36:58 crc kubenswrapper[4968]: I0218 15:36:58.319687 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" event={"ID":"a777cdac-6acd-45d2-9262-8c77577fe2b6","Type":"ContainerStarted","Data":"3faa7d11d9f728fa945a3ddb4d3a8b5f62ae326470926833b26fae43cd853289"} Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.136979 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.175130 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-catalog-content\") pod \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\" (UID: \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\") " Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.175312 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-utilities\") pod \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\" (UID: \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\") " Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.175395 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdcpc\" (UniqueName: \"kubernetes.io/projected/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-kube-api-access-cdcpc\") pod \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\" (UID: \"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d\") " Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.176195 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-utilities" (OuterVolumeSpecName: "utilities") pod "735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" (UID: "735e3828-d2af-4d0f-81d8-daf5c6ea8e4d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.182543 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-kube-api-access-cdcpc" (OuterVolumeSpecName: "kube-api-access-cdcpc") pod "735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" (UID: "735e3828-d2af-4d0f-81d8-daf5c6ea8e4d"). InnerVolumeSpecName "kube-api-access-cdcpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.230289 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" (UID: "735e3828-d2af-4d0f-81d8-daf5c6ea8e4d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.277205 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.277232 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdcpc\" (UniqueName: \"kubernetes.io/projected/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-kube-api-access-cdcpc\") on node \"crc\" DevicePath \"\"" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.277243 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.329271 4968 generic.go:334] "Generic (PLEG): container finished" podID="735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" containerID="123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc" exitCode=0 Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.329341 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh62p" event={"ID":"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d","Type":"ContainerDied","Data":"123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc"} Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.329369 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh62p" event={"ID":"735e3828-d2af-4d0f-81d8-daf5c6ea8e4d","Type":"ContainerDied","Data":"7eb6f9a688fcde99fecd57f8cb3462b17202c1ba0d02aed858cd571c82e6d259"} Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.329389 4968 scope.go:117] "RemoveContainer" containerID="123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.329413 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jh62p" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.354052 4968 generic.go:334] "Generic (PLEG): container finished" podID="a777cdac-6acd-45d2-9262-8c77577fe2b6" containerID="dfce70ed9dce24422b14311d83f4b569e33b8051116ae636b537f5cf07ec5372" exitCode=0 Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.354103 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" event={"ID":"a777cdac-6acd-45d2-9262-8c77577fe2b6","Type":"ContainerDied","Data":"dfce70ed9dce24422b14311d83f4b569e33b8051116ae636b537f5cf07ec5372"} Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.377350 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jh62p"] Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.392561 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jh62p"] Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.397251 4968 scope.go:117] "RemoveContainer" containerID="60990d97f8cabe1206fd797563fce683777c6feb43844093a225897610bebec1" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.414093 4968 scope.go:117] "RemoveContainer" containerID="495db75894ccecb54599d599c2b306f19e348a253687494714687b6d51d387c3" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.447697 4968 scope.go:117] "RemoveContainer" containerID="123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc" Feb 18 15:36:59 crc kubenswrapper[4968]: E0218 15:36:59.448970 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc\": container with ID starting with 123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc not found: ID does not exist" containerID="123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.449012 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc"} err="failed to get container status \"123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc\": rpc error: code = NotFound desc = could not find container \"123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc\": container with ID starting with 123ee21819710ac32ec17ba1ccb6b75fe90819de15dc5abb11a7636b3a852afc not found: ID does not exist" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.449040 4968 scope.go:117] "RemoveContainer" containerID="60990d97f8cabe1206fd797563fce683777c6feb43844093a225897610bebec1" Feb 18 15:36:59 crc kubenswrapper[4968]: E0218 15:36:59.449983 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60990d97f8cabe1206fd797563fce683777c6feb43844093a225897610bebec1\": container with ID starting with 60990d97f8cabe1206fd797563fce683777c6feb43844093a225897610bebec1 not found: ID does not exist" containerID="60990d97f8cabe1206fd797563fce683777c6feb43844093a225897610bebec1" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.450026 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60990d97f8cabe1206fd797563fce683777c6feb43844093a225897610bebec1"} err="failed to get container status \"60990d97f8cabe1206fd797563fce683777c6feb43844093a225897610bebec1\": rpc error: code = NotFound desc = could not find container \"60990d97f8cabe1206fd797563fce683777c6feb43844093a225897610bebec1\": container with ID starting with 60990d97f8cabe1206fd797563fce683777c6feb43844093a225897610bebec1 not found: ID does not exist" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.450055 4968 scope.go:117] "RemoveContainer" containerID="495db75894ccecb54599d599c2b306f19e348a253687494714687b6d51d387c3" Feb 18 15:36:59 crc kubenswrapper[4968]: E0218 15:36:59.450387 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"495db75894ccecb54599d599c2b306f19e348a253687494714687b6d51d387c3\": container with ID starting with 495db75894ccecb54599d599c2b306f19e348a253687494714687b6d51d387c3 not found: ID does not exist" containerID="495db75894ccecb54599d599c2b306f19e348a253687494714687b6d51d387c3" Feb 18 15:36:59 crc kubenswrapper[4968]: I0218 15:36:59.450409 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"495db75894ccecb54599d599c2b306f19e348a253687494714687b6d51d387c3"} err="failed to get container status \"495db75894ccecb54599d599c2b306f19e348a253687494714687b6d51d387c3\": rpc error: code = NotFound desc = could not find container \"495db75894ccecb54599d599c2b306f19e348a253687494714687b6d51d387c3\": container with ID starting with 495db75894ccecb54599d599c2b306f19e348a253687494714687b6d51d387c3 not found: ID does not exist" Feb 18 15:37:00 crc kubenswrapper[4968]: I0218 15:37:00.365038 4968 generic.go:334] "Generic (PLEG): container finished" podID="a777cdac-6acd-45d2-9262-8c77577fe2b6" containerID="d915a9338cca5f550a335dc44bcbdcbdbea71765ec7e283203f26253663ecce9" exitCode=0 Feb 18 15:37:00 crc kubenswrapper[4968]: I0218 15:37:00.365100 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" event={"ID":"a777cdac-6acd-45d2-9262-8c77577fe2b6","Type":"ContainerDied","Data":"d915a9338cca5f550a335dc44bcbdcbdbea71765ec7e283203f26253663ecce9"} Feb 18 15:37:01 crc kubenswrapper[4968]: I0218 15:37:01.239546 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" path="/var/lib/kubelet/pods/735e3828-d2af-4d0f-81d8-daf5c6ea8e4d/volumes" Feb 18 15:37:01 crc kubenswrapper[4968]: I0218 15:37:01.716794 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:37:01 crc kubenswrapper[4968]: I0218 15:37:01.814191 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a777cdac-6acd-45d2-9262-8c77577fe2b6-util\") pod \"a777cdac-6acd-45d2-9262-8c77577fe2b6\" (UID: \"a777cdac-6acd-45d2-9262-8c77577fe2b6\") " Feb 18 15:37:01 crc kubenswrapper[4968]: I0218 15:37:01.814280 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v845k\" (UniqueName: \"kubernetes.io/projected/a777cdac-6acd-45d2-9262-8c77577fe2b6-kube-api-access-v845k\") pod \"a777cdac-6acd-45d2-9262-8c77577fe2b6\" (UID: \"a777cdac-6acd-45d2-9262-8c77577fe2b6\") " Feb 18 15:37:01 crc kubenswrapper[4968]: I0218 15:37:01.814376 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a777cdac-6acd-45d2-9262-8c77577fe2b6-bundle\") pod \"a777cdac-6acd-45d2-9262-8c77577fe2b6\" (UID: \"a777cdac-6acd-45d2-9262-8c77577fe2b6\") " Feb 18 15:37:01 crc kubenswrapper[4968]: I0218 15:37:01.815960 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a777cdac-6acd-45d2-9262-8c77577fe2b6-bundle" (OuterVolumeSpecName: "bundle") pod "a777cdac-6acd-45d2-9262-8c77577fe2b6" (UID: "a777cdac-6acd-45d2-9262-8c77577fe2b6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:37:01 crc kubenswrapper[4968]: I0218 15:37:01.822618 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a777cdac-6acd-45d2-9262-8c77577fe2b6-kube-api-access-v845k" (OuterVolumeSpecName: "kube-api-access-v845k") pod "a777cdac-6acd-45d2-9262-8c77577fe2b6" (UID: "a777cdac-6acd-45d2-9262-8c77577fe2b6"). InnerVolumeSpecName "kube-api-access-v845k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:37:01 crc kubenswrapper[4968]: I0218 15:37:01.834615 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a777cdac-6acd-45d2-9262-8c77577fe2b6-util" (OuterVolumeSpecName: "util") pod "a777cdac-6acd-45d2-9262-8c77577fe2b6" (UID: "a777cdac-6acd-45d2-9262-8c77577fe2b6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:37:01 crc kubenswrapper[4968]: I0218 15:37:01.916063 4968 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a777cdac-6acd-45d2-9262-8c77577fe2b6-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:37:01 crc kubenswrapper[4968]: I0218 15:37:01.916489 4968 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a777cdac-6acd-45d2-9262-8c77577fe2b6-util\") on node \"crc\" DevicePath \"\"" Feb 18 15:37:01 crc kubenswrapper[4968]: I0218 15:37:01.916511 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v845k\" (UniqueName: \"kubernetes.io/projected/a777cdac-6acd-45d2-9262-8c77577fe2b6-kube-api-access-v845k\") on node \"crc\" DevicePath \"\"" Feb 18 15:37:02 crc kubenswrapper[4968]: I0218 15:37:02.377866 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" event={"ID":"a777cdac-6acd-45d2-9262-8c77577fe2b6","Type":"ContainerDied","Data":"3faa7d11d9f728fa945a3ddb4d3a8b5f62ae326470926833b26fae43cd853289"} Feb 18 15:37:02 crc kubenswrapper[4968]: I0218 15:37:02.377905 4968 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3faa7d11d9f728fa945a3ddb4d3a8b5f62ae326470926833b26fae43cd853289" Feb 18 15:37:02 crc kubenswrapper[4968]: I0218 15:37:02.377962 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.577379 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4"] Feb 18 15:37:06 crc kubenswrapper[4968]: E0218 15:37:06.577937 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a777cdac-6acd-45d2-9262-8c77577fe2b6" containerName="extract" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.577952 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="a777cdac-6acd-45d2-9262-8c77577fe2b6" containerName="extract" Feb 18 15:37:06 crc kubenswrapper[4968]: E0218 15:37:06.577965 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" containerName="registry-server" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.577974 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" containerName="registry-server" Feb 18 15:37:06 crc kubenswrapper[4968]: E0218 15:37:06.577989 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a777cdac-6acd-45d2-9262-8c77577fe2b6" containerName="util" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.577999 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="a777cdac-6acd-45d2-9262-8c77577fe2b6" containerName="util" Feb 18 15:37:06 crc kubenswrapper[4968]: E0218 15:37:06.578015 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" containerName="extract-content" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.578025 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" containerName="extract-content" Feb 18 15:37:06 crc kubenswrapper[4968]: E0218 15:37:06.578040 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a777cdac-6acd-45d2-9262-8c77577fe2b6" containerName="pull" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.578048 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="a777cdac-6acd-45d2-9262-8c77577fe2b6" containerName="pull" Feb 18 15:37:06 crc kubenswrapper[4968]: E0218 15:37:06.578060 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" containerName="extract-utilities" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.578068 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" containerName="extract-utilities" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.578206 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="a777cdac-6acd-45d2-9262-8c77577fe2b6" containerName="extract" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.578220 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="735e3828-d2af-4d0f-81d8-daf5c6ea8e4d" containerName="registry-server" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.578691 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.581131 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-zzd8r" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.608501 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4"] Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.683423 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4nbg\" (UniqueName: \"kubernetes.io/projected/40e6b6c3-28b7-4626-aff6-5b4932a65ec7-kube-api-access-q4nbg\") pod \"openstack-operator-controller-init-77859ccd85-qfms4\" (UID: \"40e6b6c3-28b7-4626-aff6-5b4932a65ec7\") " pod="openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.784966 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4nbg\" (UniqueName: \"kubernetes.io/projected/40e6b6c3-28b7-4626-aff6-5b4932a65ec7-kube-api-access-q4nbg\") pod \"openstack-operator-controller-init-77859ccd85-qfms4\" (UID: \"40e6b6c3-28b7-4626-aff6-5b4932a65ec7\") " pod="openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.806472 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4nbg\" (UniqueName: \"kubernetes.io/projected/40e6b6c3-28b7-4626-aff6-5b4932a65ec7-kube-api-access-q4nbg\") pod \"openstack-operator-controller-init-77859ccd85-qfms4\" (UID: \"40e6b6c3-28b7-4626-aff6-5b4932a65ec7\") " pod="openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4" Feb 18 15:37:06 crc kubenswrapper[4968]: I0218 15:37:06.901473 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4" Feb 18 15:37:07 crc kubenswrapper[4968]: I0218 15:37:07.136969 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4"] Feb 18 15:37:07 crc kubenswrapper[4968]: I0218 15:37:07.423057 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4" event={"ID":"40e6b6c3-28b7-4626-aff6-5b4932a65ec7","Type":"ContainerStarted","Data":"b46d98468edfd57d7fc869c39498d413d9212f0bc28803d0cbd218b39e6d45d8"} Feb 18 15:37:11 crc kubenswrapper[4968]: I0218 15:37:11.451650 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4" event={"ID":"40e6b6c3-28b7-4626-aff6-5b4932a65ec7","Type":"ContainerStarted","Data":"8507adfaca946904edeb2fae0e1b9dc7d4fa5b7a344eb4899a35a0d750956075"} Feb 18 15:37:11 crc kubenswrapper[4968]: I0218 15:37:11.452880 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4" Feb 18 15:37:11 crc kubenswrapper[4968]: I0218 15:37:11.481823 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4" podStartSLOduration=1.872369159 podStartE2EDuration="5.481807055s" podCreationTimestamp="2026-02-18 15:37:06 +0000 UTC" firstStartedPulling="2026-02-18 15:37:07.145270598 +0000 UTC m=+926.530715460" lastFinishedPulling="2026-02-18 15:37:10.754708494 +0000 UTC m=+930.140153356" observedRunningTime="2026-02-18 15:37:11.478264856 +0000 UTC m=+930.863709758" watchObservedRunningTime="2026-02-18 15:37:11.481807055 +0000 UTC m=+930.867251917" Feb 18 15:37:14 crc kubenswrapper[4968]: I0218 15:37:14.367215 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:37:14 crc kubenswrapper[4968]: I0218 15:37:14.367731 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:37:16 crc kubenswrapper[4968]: I0218 15:37:16.906688 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-77859ccd85-qfms4" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.220703 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.222276 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.226341 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-7sggr" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.242582 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.243287 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.243440 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.251106 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-ms79v" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.290771 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.291661 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.301142 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-qsjwd" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.304803 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.305607 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.311083 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-8j4kr" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.320272 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.331936 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.368941 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9tvw\" (UniqueName: \"kubernetes.io/projected/392e8c65-7168-4e8b-b672-fb0b53da6ad5-kube-api-access-n9tvw\") pod \"barbican-operator-controller-manager-868647ff47-7rw4c\" (UID: \"392e8c65-7168-4e8b-b672-fb0b53da6ad5\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.369024 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl5zs\" (UniqueName: \"kubernetes.io/projected/87fee87a-b2aa-46b7-8f11-89b00fe7638c-kube-api-access-zl5zs\") pod \"designate-operator-controller-manager-6d8bf5c495-zc86v\" (UID: \"87fee87a-b2aa-46b7-8f11-89b00fe7638c\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.369048 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcphf\" (UniqueName: \"kubernetes.io/projected/35c5d058-dc84-4faf-a645-c8463d0e6eb1-kube-api-access-vcphf\") pod \"cinder-operator-controller-manager-5d946d989d-8ng62\" (UID: \"35c5d058-dc84-4faf-a645-c8463d0e6eb1\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.369088 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-468ss\" (UniqueName: \"kubernetes.io/projected/99b546dc-24de-4e98-9fce-7d457bf981ec-kube-api-access-468ss\") pod \"glance-operator-controller-manager-77987464f4-7ltz5\" (UID: \"99b546dc-24de-4e98-9fce-7d457bf981ec\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.371856 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.388044 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.388883 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.390611 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.391336 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.393716 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-b65st" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.394006 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-mwncs" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.409817 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.425809 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.436758 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.437558 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.441390 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-92p4n" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.443220 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.453046 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.453923 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.462673 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.463546 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.464095 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-4hjmz" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.468107 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-7k68m" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.469674 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4hjk\" (UniqueName: \"kubernetes.io/projected/cbc0c461-56e5-4386-bb50-a633b89911e0-kube-api-access-h4hjk\") pod \"heat-operator-controller-manager-69f49c598c-kkzx6\" (UID: \"cbc0c461-56e5-4386-bb50-a633b89911e0\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.469798 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl5zs\" (UniqueName: \"kubernetes.io/projected/87fee87a-b2aa-46b7-8f11-89b00fe7638c-kube-api-access-zl5zs\") pod \"designate-operator-controller-manager-6d8bf5c495-zc86v\" (UID: \"87fee87a-b2aa-46b7-8f11-89b00fe7638c\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.469836 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcphf\" (UniqueName: \"kubernetes.io/projected/35c5d058-dc84-4faf-a645-c8463d0e6eb1-kube-api-access-vcphf\") pod \"cinder-operator-controller-manager-5d946d989d-8ng62\" (UID: \"35c5d058-dc84-4faf-a645-c8463d0e6eb1\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.469871 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpr2f\" (UniqueName: \"kubernetes.io/projected/cae81ff3-9955-429b-9e24-3093d3c7f1fa-kube-api-access-qpr2f\") pod \"horizon-operator-controller-manager-5b9b8895d5-v9gsm\" (UID: \"cae81ff3-9955-429b-9e24-3093d3c7f1fa\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.469903 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-468ss\" (UniqueName: \"kubernetes.io/projected/99b546dc-24de-4e98-9fce-7d457bf981ec-kube-api-access-468ss\") pod \"glance-operator-controller-manager-77987464f4-7ltz5\" (UID: \"99b546dc-24de-4e98-9fce-7d457bf981ec\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.469930 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9tvw\" (UniqueName: \"kubernetes.io/projected/392e8c65-7168-4e8b-b672-fb0b53da6ad5-kube-api-access-n9tvw\") pod \"barbican-operator-controller-manager-868647ff47-7rw4c\" (UID: \"392e8c65-7168-4e8b-b672-fb0b53da6ad5\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.476808 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.477585 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.483129 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.485205 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-9mkkv" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.488703 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.510057 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcphf\" (UniqueName: \"kubernetes.io/projected/35c5d058-dc84-4faf-a645-c8463d0e6eb1-kube-api-access-vcphf\") pod \"cinder-operator-controller-manager-5d946d989d-8ng62\" (UID: \"35c5d058-dc84-4faf-a645-c8463d0e6eb1\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.510502 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9tvw\" (UniqueName: \"kubernetes.io/projected/392e8c65-7168-4e8b-b672-fb0b53da6ad5-kube-api-access-n9tvw\") pod \"barbican-operator-controller-manager-868647ff47-7rw4c\" (UID: \"392e8c65-7168-4e8b-b672-fb0b53da6ad5\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.512839 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-468ss\" (UniqueName: \"kubernetes.io/projected/99b546dc-24de-4e98-9fce-7d457bf981ec-kube-api-access-468ss\") pod \"glance-operator-controller-manager-77987464f4-7ltz5\" (UID: \"99b546dc-24de-4e98-9fce-7d457bf981ec\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.521389 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl5zs\" (UniqueName: \"kubernetes.io/projected/87fee87a-b2aa-46b7-8f11-89b00fe7638c-kube-api-access-zl5zs\") pod \"designate-operator-controller-manager-6d8bf5c495-zc86v\" (UID: \"87fee87a-b2aa-46b7-8f11-89b00fe7638c\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.547213 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.558111 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.558965 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.562203 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-5m9lt" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.574640 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.575455 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert\") pod \"infra-operator-controller-manager-79d975b745-lwn4x\" (UID: \"fb1b11dc-0a5b-4784-bc3b-a0976c481831\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.575600 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpr2f\" (UniqueName: \"kubernetes.io/projected/cae81ff3-9955-429b-9e24-3093d3c7f1fa-kube-api-access-qpr2f\") pod \"horizon-operator-controller-manager-5b9b8895d5-v9gsm\" (UID: \"cae81ff3-9955-429b-9e24-3093d3c7f1fa\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.575683 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqsnm\" (UniqueName: \"kubernetes.io/projected/debbb308-3378-47b6-8e2a-b0d8edc95bfc-kube-api-access-cqsnm\") pod \"manila-operator-controller-manager-54f6768c69-f4prf\" (UID: \"debbb308-3378-47b6-8e2a-b0d8edc95bfc\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.575781 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j78bp\" (UniqueName: \"kubernetes.io/projected/fb1b11dc-0a5b-4784-bc3b-a0976c481831-kube-api-access-j78bp\") pod \"infra-operator-controller-manager-79d975b745-lwn4x\" (UID: \"fb1b11dc-0a5b-4784-bc3b-a0976c481831\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.575858 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8gl8\" (UniqueName: \"kubernetes.io/projected/60510787-a5c0-467a-969f-38c6be9b2184-kube-api-access-c8gl8\") pod \"keystone-operator-controller-manager-b4d948c87-6jkdl\" (UID: \"60510787-a5c0-467a-969f-38c6be9b2184\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.576029 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4hjk\" (UniqueName: \"kubernetes.io/projected/cbc0c461-56e5-4386-bb50-a633b89911e0-kube-api-access-h4hjk\") pod \"heat-operator-controller-manager-69f49c598c-kkzx6\" (UID: \"cbc0c461-56e5-4386-bb50-a633b89911e0\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.576099 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpzql\" (UniqueName: \"kubernetes.io/projected/6f7c68fd-afbb-4751-b1da-d27bfe32d6c7-kube-api-access-jpzql\") pod \"ironic-operator-controller-manager-554564d7fc-56xzj\" (UID: \"6f7c68fd-afbb-4751-b1da-d27bfe32d6c7\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.586076 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.613658 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpr2f\" (UniqueName: \"kubernetes.io/projected/cae81ff3-9955-429b-9e24-3093d3c7f1fa-kube-api-access-qpr2f\") pod \"horizon-operator-controller-manager-5b9b8895d5-v9gsm\" (UID: \"cae81ff3-9955-429b-9e24-3093d3c7f1fa\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.635163 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.638067 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4hjk\" (UniqueName: \"kubernetes.io/projected/cbc0c461-56e5-4386-bb50-a633b89911e0-kube-api-access-h4hjk\") pod \"heat-operator-controller-manager-69f49c598c-kkzx6\" (UID: \"cbc0c461-56e5-4386-bb50-a633b89911e0\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.639167 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.669347 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.672293 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.673229 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.678108 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqsnm\" (UniqueName: \"kubernetes.io/projected/debbb308-3378-47b6-8e2a-b0d8edc95bfc-kube-api-access-cqsnm\") pod \"manila-operator-controller-manager-54f6768c69-f4prf\" (UID: \"debbb308-3378-47b6-8e2a-b0d8edc95bfc\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.678147 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j78bp\" (UniqueName: \"kubernetes.io/projected/fb1b11dc-0a5b-4784-bc3b-a0976c481831-kube-api-access-j78bp\") pod \"infra-operator-controller-manager-79d975b745-lwn4x\" (UID: \"fb1b11dc-0a5b-4784-bc3b-a0976c481831\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.678165 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8gl8\" (UniqueName: \"kubernetes.io/projected/60510787-a5c0-467a-969f-38c6be9b2184-kube-api-access-c8gl8\") pod \"keystone-operator-controller-manager-b4d948c87-6jkdl\" (UID: \"60510787-a5c0-467a-969f-38c6be9b2184\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.678189 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smsvf\" (UniqueName: \"kubernetes.io/projected/00926e2b-9085-412d-9b65-2c71f76bd9ed-kube-api-access-smsvf\") pod \"mariadb-operator-controller-manager-6994f66f48-mbn84\" (UID: \"00926e2b-9085-412d-9b65-2c71f76bd9ed\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.678228 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpzql\" (UniqueName: \"kubernetes.io/projected/6f7c68fd-afbb-4751-b1da-d27bfe32d6c7-kube-api-access-jpzql\") pod \"ironic-operator-controller-manager-554564d7fc-56xzj\" (UID: \"6f7c68fd-afbb-4751-b1da-d27bfe32d6c7\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.678256 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert\") pod \"infra-operator-controller-manager-79d975b745-lwn4x\" (UID: \"fb1b11dc-0a5b-4784-bc3b-a0976c481831\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:37:37 crc kubenswrapper[4968]: E0218 15:37:37.678366 4968 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 18 15:37:37 crc kubenswrapper[4968]: E0218 15:37:37.678413 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert podName:fb1b11dc-0a5b-4784-bc3b-a0976c481831 nodeName:}" failed. No retries permitted until 2026-02-18 15:37:38.17839597 +0000 UTC m=+957.563840832 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert") pod "infra-operator-controller-manager-79d975b745-lwn4x" (UID: "fb1b11dc-0a5b-4784-bc3b-a0976c481831") : secret "infra-operator-webhook-server-cert" not found Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.681189 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.685195 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-fzfb8" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.705277 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqsnm\" (UniqueName: \"kubernetes.io/projected/debbb308-3378-47b6-8e2a-b0d8edc95bfc-kube-api-access-cqsnm\") pod \"manila-operator-controller-manager-54f6768c69-f4prf\" (UID: \"debbb308-3378-47b6-8e2a-b0d8edc95bfc\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.707939 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8gl8\" (UniqueName: \"kubernetes.io/projected/60510787-a5c0-467a-969f-38c6be9b2184-kube-api-access-c8gl8\") pod \"keystone-operator-controller-manager-b4d948c87-6jkdl\" (UID: \"60510787-a5c0-467a-969f-38c6be9b2184\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.712690 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j78bp\" (UniqueName: \"kubernetes.io/projected/fb1b11dc-0a5b-4784-bc3b-a0976c481831-kube-api-access-j78bp\") pod \"infra-operator-controller-manager-79d975b745-lwn4x\" (UID: \"fb1b11dc-0a5b-4784-bc3b-a0976c481831\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.713687 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpzql\" (UniqueName: \"kubernetes.io/projected/6f7c68fd-afbb-4751-b1da-d27bfe32d6c7-kube-api-access-jpzql\") pod \"ironic-operator-controller-manager-554564d7fc-56xzj\" (UID: \"6f7c68fd-afbb-4751-b1da-d27bfe32d6c7\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.715249 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.724026 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.725223 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.732862 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-fwt48" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.733307 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.739150 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.756956 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.762555 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.763343 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.767372 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-xdjgq" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.774664 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.775725 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.777380 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-7587k" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.780885 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smsvf\" (UniqueName: \"kubernetes.io/projected/00926e2b-9085-412d-9b65-2c71f76bd9ed-kube-api-access-smsvf\") pod \"mariadb-operator-controller-manager-6994f66f48-mbn84\" (UID: \"00926e2b-9085-412d-9b65-2c71f76bd9ed\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.780961 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b5ht\" (UniqueName: \"kubernetes.io/projected/1bee3920-4de9-436c-8d80-4ea7f0903f11-kube-api-access-7b5ht\") pod \"neutron-operator-controller-manager-64ddbf8bb-vvqqh\" (UID: \"1bee3920-4de9-436c-8d80-4ea7f0903f11\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.780990 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzrjj\" (UniqueName: \"kubernetes.io/projected/aa54cd8f-435a-4055-9668-1ab7c3e72516-kube-api-access-qzrjj\") pod \"nova-operator-controller-manager-567668f5cf-xhwzj\" (UID: \"aa54cd8f-435a-4055-9668-1ab7c3e72516\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.781493 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.787113 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.788870 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.789948 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.793954 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-dmhbs" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.794651 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.796765 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.802235 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.811612 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.820267 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.821178 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.824667 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smsvf\" (UniqueName: \"kubernetes.io/projected/00926e2b-9085-412d-9b65-2c71f76bd9ed-kube-api-access-smsvf\") pod \"mariadb-operator-controller-manager-6994f66f48-mbn84\" (UID: \"00926e2b-9085-412d-9b65-2c71f76bd9ed\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.828614 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-d8r8b" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.836147 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.837549 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.847791 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.851593 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-g9jm9" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.862353 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.873488 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.874779 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.876625 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-c6d2b" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.883014 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x5nw\" (UniqueName: \"kubernetes.io/projected/53297588-6688-444d-952c-810b9b265acf-kube-api-access-8x5nw\") pod \"ovn-operator-controller-manager-d44cf6b75-jhq7j\" (UID: \"53297588-6688-444d-952c-810b9b265acf\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.883059 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz\" (UID: \"c2c93869-c328-4b28-8098-18191433d4d2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.883088 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq5lh\" (UniqueName: \"kubernetes.io/projected/c2c93869-c328-4b28-8098-18191433d4d2-kube-api-access-hq5lh\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz\" (UID: \"c2c93869-c328-4b28-8098-18191433d4d2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.883107 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5pdc\" (UniqueName: \"kubernetes.io/projected/e39cd00d-1815-442a-9343-277280cf8ccc-kube-api-access-c5pdc\") pod \"swift-operator-controller-manager-68f46476f-n6bn4\" (UID: \"e39cd00d-1815-442a-9343-277280cf8ccc\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.883129 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b5ht\" (UniqueName: \"kubernetes.io/projected/1bee3920-4de9-436c-8d80-4ea7f0903f11-kube-api-access-7b5ht\") pod \"neutron-operator-controller-manager-64ddbf8bb-vvqqh\" (UID: \"1bee3920-4de9-436c-8d80-4ea7f0903f11\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.883149 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9qjq\" (UniqueName: \"kubernetes.io/projected/8a5aaad5-3e01-4de2-9bf1-e64680ce7b18-kube-api-access-s9qjq\") pod \"octavia-operator-controller-manager-69f8888797-dfg2t\" (UID: \"8a5aaad5-3e01-4de2-9bf1-e64680ce7b18\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.883166 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzrjj\" (UniqueName: \"kubernetes.io/projected/aa54cd8f-435a-4055-9668-1ab7c3e72516-kube-api-access-qzrjj\") pod \"nova-operator-controller-manager-567668f5cf-xhwzj\" (UID: \"aa54cd8f-435a-4055-9668-1ab7c3e72516\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.883201 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzpxg\" (UniqueName: \"kubernetes.io/projected/964d8817-99e3-4a8d-91ab-876ec187f62d-kube-api-access-bzpxg\") pod \"placement-operator-controller-manager-8497b45c89-wbp4k\" (UID: \"964d8817-99e3-4a8d-91ab-876ec187f62d\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.896526 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.905509 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b5ht\" (UniqueName: \"kubernetes.io/projected/1bee3920-4de9-436c-8d80-4ea7f0903f11-kube-api-access-7b5ht\") pod \"neutron-operator-controller-manager-64ddbf8bb-vvqqh\" (UID: \"1bee3920-4de9-436c-8d80-4ea7f0903f11\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.911560 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzrjj\" (UniqueName: \"kubernetes.io/projected/aa54cd8f-435a-4055-9668-1ab7c3e72516-kube-api-access-qzrjj\") pod \"nova-operator-controller-manager-567668f5cf-xhwzj\" (UID: \"aa54cd8f-435a-4055-9668-1ab7c3e72516\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.926093 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-qphmd"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.927167 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.929558 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-8pcpd" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.935025 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.943274 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-qphmd"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.967384 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.983868 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz"] Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.984808 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.985015 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x5nw\" (UniqueName: \"kubernetes.io/projected/53297588-6688-444d-952c-810b9b265acf-kube-api-access-8x5nw\") pod \"ovn-operator-controller-manager-d44cf6b75-jhq7j\" (UID: \"53297588-6688-444d-952c-810b9b265acf\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.985055 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz\" (UID: \"c2c93869-c328-4b28-8098-18191433d4d2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.985074 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq5lh\" (UniqueName: \"kubernetes.io/projected/c2c93869-c328-4b28-8098-18191433d4d2-kube-api-access-hq5lh\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz\" (UID: \"c2c93869-c328-4b28-8098-18191433d4d2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.985094 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5pdc\" (UniqueName: \"kubernetes.io/projected/e39cd00d-1815-442a-9343-277280cf8ccc-kube-api-access-c5pdc\") pod \"swift-operator-controller-manager-68f46476f-n6bn4\" (UID: \"e39cd00d-1815-442a-9343-277280cf8ccc\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.985116 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9qjq\" (UniqueName: \"kubernetes.io/projected/8a5aaad5-3e01-4de2-9bf1-e64680ce7b18-kube-api-access-s9qjq\") pod \"octavia-operator-controller-manager-69f8888797-dfg2t\" (UID: \"8a5aaad5-3e01-4de2-9bf1-e64680ce7b18\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.985163 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gcjk\" (UniqueName: \"kubernetes.io/projected/095020a1-2516-42f1-9223-339fd943c52f-kube-api-access-7gcjk\") pod \"telemetry-operator-controller-manager-7f45b4ff68-kskjh\" (UID: \"095020a1-2516-42f1-9223-339fd943c52f\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.985182 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzpxg\" (UniqueName: \"kubernetes.io/projected/964d8817-99e3-4a8d-91ab-876ec187f62d-kube-api-access-bzpxg\") pod \"placement-operator-controller-manager-8497b45c89-wbp4k\" (UID: \"964d8817-99e3-4a8d-91ab-876ec187f62d\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.985198 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zplxr\" (UniqueName: \"kubernetes.io/projected/dc5fcae4-29ef-4706-9b51-e4f3e99f1c36-kube-api-access-zplxr\") pod \"test-operator-controller-manager-7866795846-qphmd\" (UID: \"dc5fcae4-29ef-4706-9b51-e4f3e99f1c36\") " pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" Feb 18 15:37:37 crc kubenswrapper[4968]: E0218 15:37:37.985529 4968 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 15:37:37 crc kubenswrapper[4968]: E0218 15:37:37.985579 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert podName:c2c93869-c328-4b28-8098-18191433d4d2 nodeName:}" failed. No retries permitted until 2026-02-18 15:37:38.485564204 +0000 UTC m=+957.871009066 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" (UID: "c2c93869-c328-4b28-8098-18191433d4d2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.994020 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-sg2b5" Feb 18 15:37:37 crc kubenswrapper[4968]: I0218 15:37:37.994177 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz"] Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.010359 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzpxg\" (UniqueName: \"kubernetes.io/projected/964d8817-99e3-4a8d-91ab-876ec187f62d-kube-api-access-bzpxg\") pod \"placement-operator-controller-manager-8497b45c89-wbp4k\" (UID: \"964d8817-99e3-4a8d-91ab-876ec187f62d\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.010643 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x5nw\" (UniqueName: \"kubernetes.io/projected/53297588-6688-444d-952c-810b9b265acf-kube-api-access-8x5nw\") pod \"ovn-operator-controller-manager-d44cf6b75-jhq7j\" (UID: \"53297588-6688-444d-952c-810b9b265acf\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.017418 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.035300 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq5lh\" (UniqueName: \"kubernetes.io/projected/c2c93869-c328-4b28-8098-18191433d4d2-kube-api-access-hq5lh\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz\" (UID: \"c2c93869-c328-4b28-8098-18191433d4d2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.047924 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.065650 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9qjq\" (UniqueName: \"kubernetes.io/projected/8a5aaad5-3e01-4de2-9bf1-e64680ce7b18-kube-api-access-s9qjq\") pod \"octavia-operator-controller-manager-69f8888797-dfg2t\" (UID: \"8a5aaad5-3e01-4de2-9bf1-e64680ce7b18\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.070528 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5pdc\" (UniqueName: \"kubernetes.io/projected/e39cd00d-1815-442a-9343-277280cf8ccc-kube-api-access-c5pdc\") pod \"swift-operator-controller-manager-68f46476f-n6bn4\" (UID: \"e39cd00d-1815-442a-9343-277280cf8ccc\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.086334 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4t8m\" (UniqueName: \"kubernetes.io/projected/56f5f54a-9f54-4e6e-a9fc-129398f0df0b-kube-api-access-b4t8m\") pod \"watcher-operator-controller-manager-5c48998cbf-jdwxz\" (UID: \"56f5f54a-9f54-4e6e-a9fc-129398f0df0b\") " pod="openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.086383 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gcjk\" (UniqueName: \"kubernetes.io/projected/095020a1-2516-42f1-9223-339fd943c52f-kube-api-access-7gcjk\") pod \"telemetry-operator-controller-manager-7f45b4ff68-kskjh\" (UID: \"095020a1-2516-42f1-9223-339fd943c52f\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.086406 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zplxr\" (UniqueName: \"kubernetes.io/projected/dc5fcae4-29ef-4706-9b51-e4f3e99f1c36-kube-api-access-zplxr\") pod \"test-operator-controller-manager-7866795846-qphmd\" (UID: \"dc5fcae4-29ef-4706-9b51-e4f3e99f1c36\") " pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.120089 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gcjk\" (UniqueName: \"kubernetes.io/projected/095020a1-2516-42f1-9223-339fd943c52f-kube-api-access-7gcjk\") pod \"telemetry-operator-controller-manager-7f45b4ff68-kskjh\" (UID: \"095020a1-2516-42f1-9223-339fd943c52f\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.140807 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g"] Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.145047 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.145836 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zplxr\" (UniqueName: \"kubernetes.io/projected/dc5fcae4-29ef-4706-9b51-e4f3e99f1c36-kube-api-access-zplxr\") pod \"test-operator-controller-manager-7866795846-qphmd\" (UID: \"dc5fcae4-29ef-4706-9b51-e4f3e99f1c36\") " pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.147599 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.147803 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.147921 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-r462p" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.147990 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.190238 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4t8m\" (UniqueName: \"kubernetes.io/projected/56f5f54a-9f54-4e6e-a9fc-129398f0df0b-kube-api-access-b4t8m\") pod \"watcher-operator-controller-manager-5c48998cbf-jdwxz\" (UID: \"56f5f54a-9f54-4e6e-a9fc-129398f0df0b\") " pod="openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.190304 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert\") pod \"infra-operator-controller-manager-79d975b745-lwn4x\" (UID: \"fb1b11dc-0a5b-4784-bc3b-a0976c481831\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:37:38 crc kubenswrapper[4968]: E0218 15:37:38.190441 4968 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 18 15:37:38 crc kubenswrapper[4968]: E0218 15:37:38.190495 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert podName:fb1b11dc-0a5b-4784-bc3b-a0976c481831 nodeName:}" failed. No retries permitted until 2026-02-18 15:37:39.190473196 +0000 UTC m=+958.575918058 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert") pod "infra-operator-controller-manager-79d975b745-lwn4x" (UID: "fb1b11dc-0a5b-4784-bc3b-a0976c481831") : secret "infra-operator-webhook-server-cert" not found Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.196542 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g"] Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.255313 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4t8m\" (UniqueName: \"kubernetes.io/projected/56f5f54a-9f54-4e6e-a9fc-129398f0df0b-kube-api-access-b4t8m\") pod \"watcher-operator-controller-manager-5c48998cbf-jdwxz\" (UID: \"56f5f54a-9f54-4e6e-a9fc-129398f0df0b\") " pod="openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.268532 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg"] Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.270728 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.273926 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-rhhgn" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.276010 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg"] Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.291548 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.291932 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.291951 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74qm2\" (UniqueName: \"kubernetes.io/projected/5e48a888-178c-4d24-b960-60b6251940db-kube-api-access-74qm2\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.300124 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.314104 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.323955 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.363992 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.390069 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.393154 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm9gb\" (UniqueName: \"kubernetes.io/projected/db247d6e-0199-4fde-85d9-86b08817a070-kube-api-access-lm9gb\") pod \"rabbitmq-cluster-operator-manager-668c99d594-98jxg\" (UID: \"db247d6e-0199-4fde-85d9-86b08817a070\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.393228 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.393342 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.393365 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74qm2\" (UniqueName: \"kubernetes.io/projected/5e48a888-178c-4d24-b960-60b6251940db-kube-api-access-74qm2\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:38 crc kubenswrapper[4968]: E0218 15:37:38.393385 4968 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 18 15:37:38 crc kubenswrapper[4968]: E0218 15:37:38.393457 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs podName:5e48a888-178c-4d24-b960-60b6251940db nodeName:}" failed. No retries permitted until 2026-02-18 15:37:38.893438034 +0000 UTC m=+958.278882896 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs") pod "openstack-operator-controller-manager-79f9cfbcf7-bzx4g" (UID: "5e48a888-178c-4d24-b960-60b6251940db") : secret "webhook-server-cert" not found Feb 18 15:37:38 crc kubenswrapper[4968]: E0218 15:37:38.393525 4968 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 18 15:37:38 crc kubenswrapper[4968]: E0218 15:37:38.393639 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs podName:5e48a888-178c-4d24-b960-60b6251940db nodeName:}" failed. No retries permitted until 2026-02-18 15:37:38.893596099 +0000 UTC m=+958.279041081 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs") pod "openstack-operator-controller-manager-79f9cfbcf7-bzx4g" (UID: "5e48a888-178c-4d24-b960-60b6251940db") : secret "metrics-server-cert" not found Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.394353 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62"] Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.424825 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74qm2\" (UniqueName: \"kubernetes.io/projected/5e48a888-178c-4d24-b960-60b6251940db-kube-api-access-74qm2\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.438523 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c"] Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.460405 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.494703 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz\" (UID: \"c2c93869-c328-4b28-8098-18191433d4d2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.494836 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm9gb\" (UniqueName: \"kubernetes.io/projected/db247d6e-0199-4fde-85d9-86b08817a070-kube-api-access-lm9gb\") pod \"rabbitmq-cluster-operator-manager-668c99d594-98jxg\" (UID: \"db247d6e-0199-4fde-85d9-86b08817a070\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg" Feb 18 15:37:38 crc kubenswrapper[4968]: E0218 15:37:38.495212 4968 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 15:37:38 crc kubenswrapper[4968]: E0218 15:37:38.495869 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert podName:c2c93869-c328-4b28-8098-18191433d4d2 nodeName:}" failed. No retries permitted until 2026-02-18 15:37:39.49585495 +0000 UTC m=+958.881299812 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" (UID: "c2c93869-c328-4b28-8098-18191433d4d2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.511394 4968 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.517729 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm9gb\" (UniqueName: \"kubernetes.io/projected/db247d6e-0199-4fde-85d9-86b08817a070-kube-api-access-lm9gb\") pod \"rabbitmq-cluster-operator-manager-668c99d594-98jxg\" (UID: \"db247d6e-0199-4fde-85d9-86b08817a070\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.649796 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c" event={"ID":"392e8c65-7168-4e8b-b672-fb0b53da6ad5","Type":"ContainerStarted","Data":"bdfa69540744ba9d33985cab4573a1d7c785419e4e7cb0e4e01fa88f5120e3a3"} Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.654143 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62" event={"ID":"35c5d058-dc84-4faf-a645-c8463d0e6eb1","Type":"ContainerStarted","Data":"f9c6dccb906df51a73724f254e8babd56090454d7f81fb4fb371dbd35ba85cc9"} Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.793288 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg" Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.899595 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:38 crc kubenswrapper[4968]: E0218 15:37:38.899933 4968 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.899991 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:38 crc kubenswrapper[4968]: E0218 15:37:38.900022 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs podName:5e48a888-178c-4d24-b960-60b6251940db nodeName:}" failed. No retries permitted until 2026-02-18 15:37:39.900000436 +0000 UTC m=+959.285445388 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs") pod "openstack-operator-controller-manager-79f9cfbcf7-bzx4g" (UID: "5e48a888-178c-4d24-b960-60b6251940db") : secret "metrics-server-cert" not found Feb 18 15:37:38 crc kubenswrapper[4968]: E0218 15:37:38.900121 4968 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 18 15:37:38 crc kubenswrapper[4968]: E0218 15:37:38.900172 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs podName:5e48a888-178c-4d24-b960-60b6251940db nodeName:}" failed. No retries permitted until 2026-02-18 15:37:39.90015777 +0000 UTC m=+959.285602632 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs") pod "openstack-operator-controller-manager-79f9cfbcf7-bzx4g" (UID: "5e48a888-178c-4d24-b960-60b6251940db") : secret "webhook-server-cert" not found Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.949836 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5"] Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.982546 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl"] Feb 18 15:37:38 crc kubenswrapper[4968]: I0218 15:37:38.997846 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v"] Feb 18 15:37:39 crc kubenswrapper[4968]: W0218 15:37:39.000228 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60510787_a5c0_467a_969f_38c6be9b2184.slice/crio-12149d0d672fe1aae2406409b1fbd70287ed5f56f33c2eed3d10f13957d07117 WatchSource:0}: Error finding container 12149d0d672fe1aae2406409b1fbd70287ed5f56f33c2eed3d10f13957d07117: Status 404 returned error can't find the container with id 12149d0d672fe1aae2406409b1fbd70287ed5f56f33c2eed3d10f13957d07117 Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.018655 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm"] Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.026529 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6"] Feb 18 15:37:39 crc kubenswrapper[4968]: W0218 15:37:39.043517 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddebbb308_3378_47b6_8e2a_b0d8edc95bfc.slice/crio-84320c1c16d1d60552e817c61776232a883e1a19e25a21b0f815ae30e4380045 WatchSource:0}: Error finding container 84320c1c16d1d60552e817c61776232a883e1a19e25a21b0f815ae30e4380045: Status 404 returned error can't find the container with id 84320c1c16d1d60552e817c61776232a883e1a19e25a21b0f815ae30e4380045 Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.051618 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84"] Feb 18 15:37:39 crc kubenswrapper[4968]: W0218 15:37:39.054929 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00926e2b_9085_412d_9b65_2c71f76bd9ed.slice/crio-4f2f98ed637845cf906d01d7ecce9c1fba4aa8e4a7a6e640acb383c38328d1f0 WatchSource:0}: Error finding container 4f2f98ed637845cf906d01d7ecce9c1fba4aa8e4a7a6e640acb383c38328d1f0: Status 404 returned error can't find the container with id 4f2f98ed637845cf906d01d7ecce9c1fba4aa8e4a7a6e640acb383c38328d1f0 Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.056888 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf"] Feb 18 15:37:39 crc kubenswrapper[4968]: W0218 15:37:39.059644 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f7c68fd_afbb_4751_b1da_d27bfe32d6c7.slice/crio-fd85b5369c26b43d86a41490d99ca4ec56e33f5301dc7af370b6f2bc8a376aa6 WatchSource:0}: Error finding container fd85b5369c26b43d86a41490d99ca4ec56e33f5301dc7af370b6f2bc8a376aa6: Status 404 returned error can't find the container with id fd85b5369c26b43d86a41490d99ca4ec56e33f5301dc7af370b6f2bc8a376aa6 Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.068021 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj"] Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.203947 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert\") pod \"infra-operator-controller-manager-79d975b745-lwn4x\" (UID: \"fb1b11dc-0a5b-4784-bc3b-a0976c481831\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.204132 4968 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.204179 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert podName:fb1b11dc-0a5b-4784-bc3b-a0976c481831 nodeName:}" failed. No retries permitted until 2026-02-18 15:37:41.204164955 +0000 UTC m=+960.589609817 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert") pod "infra-operator-controller-manager-79d975b745-lwn4x" (UID: "fb1b11dc-0a5b-4784-bc3b-a0976c481831") : secret "infra-operator-webhook-server-cert" not found Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.408696 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj"] Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.415542 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t"] Feb 18 15:37:39 crc kubenswrapper[4968]: W0218 15:37:39.425163 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a5aaad5_3e01_4de2_9bf1_e64680ce7b18.slice/crio-b21e2d8f645eb88f04cdac395752d7bed65436bcc624a6dc228ad8c119e6ef06 WatchSource:0}: Error finding container b21e2d8f645eb88f04cdac395752d7bed65436bcc624a6dc228ad8c119e6ef06: Status 404 returned error can't find the container with id b21e2d8f645eb88f04cdac395752d7bed65436bcc624a6dc228ad8c119e6ef06 Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.434041 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz"] Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.448662 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh"] Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.455773 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4"] Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.466522 4968 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bzpxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-8497b45c89-wbp4k_openstack-operators(964d8817-99e3-4a8d-91ab-876ec187f62d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.467041 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k"] Feb 18 15:37:39 crc kubenswrapper[4968]: W0218 15:37:39.467357 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc5fcae4_29ef_4706_9b51_e4f3e99f1c36.slice/crio-212fa8311de9d3b8c3a25af06b2aa731a51c713c0f4ec62061fda456170d1c8e WatchSource:0}: Error finding container 212fa8311de9d3b8c3a25af06b2aa731a51c713c0f4ec62061fda456170d1c8e: Status 404 returned error can't find the container with id 212fa8311de9d3b8c3a25af06b2aa731a51c713c0f4ec62061fda456170d1c8e Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.467846 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" podUID="964d8817-99e3-4a8d-91ab-876ec187f62d" Feb 18 15:37:39 crc kubenswrapper[4968]: W0218 15:37:39.468939 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53297588_6688_444d_952c_810b9b265acf.slice/crio-d33d199b30822c15b45b551d6935ecbad9e92748f59e0b0816c96a9ceb8ae8bc WatchSource:0}: Error finding container d33d199b30822c15b45b551d6935ecbad9e92748f59e0b0816c96a9ceb8ae8bc: Status 404 returned error can't find the container with id d33d199b30822c15b45b551d6935ecbad9e92748f59e0b0816c96a9ceb8ae8bc Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.471348 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg"] Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.471797 4968 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8x5nw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-d44cf6b75-jhq7j_openstack-operators(53297588-6688-444d-952c-810b9b265acf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.473602 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" podUID="53297588-6688-444d-952c-810b9b265acf" Feb 18 15:37:39 crc kubenswrapper[4968]: W0218 15:37:39.473827 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod095020a1_2516_42f1_9223_339fd943c52f.slice/crio-39bc0df625306498312f3879ac4b28fecd8b3a5ff303082dcb721c5869298968 WatchSource:0}: Error finding container 39bc0df625306498312f3879ac4b28fecd8b3a5ff303082dcb721c5869298968: Status 404 returned error can't find the container with id 39bc0df625306498312f3879ac4b28fecd8b3a5ff303082dcb721c5869298968 Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.476232 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j"] Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.478574 4968 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zplxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-7866795846-qphmd_openstack-operators(dc5fcae4-29ef-4706-9b51-e4f3e99f1c36): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.479790 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" podUID="dc5fcae4-29ef-4706-9b51-e4f3e99f1c36" Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.485587 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-qphmd"] Feb 18 15:37:39 crc kubenswrapper[4968]: W0218 15:37:39.486888 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode39cd00d_1815_442a_9343_277280cf8ccc.slice/crio-275b82a5615e1d1c5032154272342040bf40b42ce1619baf78a0d1796c356628 WatchSource:0}: Error finding container 275b82a5615e1d1c5032154272342040bf40b42ce1619baf78a0d1796c356628: Status 404 returned error can't find the container with id 275b82a5615e1d1c5032154272342040bf40b42ce1619baf78a0d1796c356628 Feb 18 15:37:39 crc kubenswrapper[4968]: W0218 15:37:39.487581 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb247d6e_0199_4fde_85d9_86b08817a070.slice/crio-0d2a9d6a1537ed6afb6b28b706ec6bd7a7cff3f280a01827c713167abdd234d6 WatchSource:0}: Error finding container 0d2a9d6a1537ed6afb6b28b706ec6bd7a7cff3f280a01827c713167abdd234d6: Status 404 returned error can't find the container with id 0d2a9d6a1537ed6afb6b28b706ec6bd7a7cff3f280a01827c713167abdd234d6 Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.487729 4968 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7gcjk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-7f45b4ff68-kskjh_openstack-operators(095020a1-2516-42f1-9223-339fd943c52f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.488976 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" podUID="095020a1-2516-42f1-9223-339fd943c52f" Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.498488 4968 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lm9gb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-98jxg_openstack-operators(db247d6e-0199-4fde-85d9-86b08817a070): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.499539 4968 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c5pdc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68f46476f-n6bn4_openstack-operators(e39cd00d-1815-442a-9343-277280cf8ccc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.499640 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg" podUID="db247d6e-0199-4fde-85d9-86b08817a070" Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.500667 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" podUID="e39cd00d-1815-442a-9343-277280cf8ccc" Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.501538 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh"] Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.515434 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz\" (UID: \"c2c93869-c328-4b28-8098-18191433d4d2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.515560 4968 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.515599 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert podName:c2c93869-c328-4b28-8098-18191433d4d2 nodeName:}" failed. No retries permitted until 2026-02-18 15:37:41.515587057 +0000 UTC m=+960.901031919 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" (UID: "c2c93869-c328-4b28-8098-18191433d4d2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.662541 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" event={"ID":"964d8817-99e3-4a8d-91ab-876ec187f62d","Type":"ContainerStarted","Data":"2626e7b9f46317f43e153dd4888e97ac10258ad8998a39166f3998866383b25f"} Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.663969 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" podUID="964d8817-99e3-4a8d-91ab-876ec187f62d" Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.665604 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" event={"ID":"53297588-6688-444d-952c-810b9b265acf","Type":"ContainerStarted","Data":"d33d199b30822c15b45b551d6935ecbad9e92748f59e0b0816c96a9ceb8ae8bc"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.667306 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t" event={"ID":"8a5aaad5-3e01-4de2-9bf1-e64680ce7b18","Type":"ContainerStarted","Data":"b21e2d8f645eb88f04cdac395752d7bed65436bcc624a6dc228ad8c119e6ef06"} Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.667782 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" podUID="53297588-6688-444d-952c-810b9b265acf" Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.668945 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" event={"ID":"e39cd00d-1815-442a-9343-277280cf8ccc","Type":"ContainerStarted","Data":"275b82a5615e1d1c5032154272342040bf40b42ce1619baf78a0d1796c356628"} Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.669921 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" podUID="e39cd00d-1815-442a-9343-277280cf8ccc" Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.671458 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj" event={"ID":"6f7c68fd-afbb-4751-b1da-d27bfe32d6c7","Type":"ContainerStarted","Data":"fd85b5369c26b43d86a41490d99ca4ec56e33f5301dc7af370b6f2bc8a376aa6"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.672526 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj" event={"ID":"aa54cd8f-435a-4055-9668-1ab7c3e72516","Type":"ContainerStarted","Data":"ee4841160e552f3abf1156b8761c63b54940289b307b987fe2d5845a347de212"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.673818 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v" event={"ID":"87fee87a-b2aa-46b7-8f11-89b00fe7638c","Type":"ContainerStarted","Data":"547a3284e212473805c078390aadb678e9b893e56e9852ed454e840b74e5862a"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.675521 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf" event={"ID":"debbb308-3378-47b6-8e2a-b0d8edc95bfc","Type":"ContainerStarted","Data":"84320c1c16d1d60552e817c61776232a883e1a19e25a21b0f815ae30e4380045"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.677307 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg" event={"ID":"db247d6e-0199-4fde-85d9-86b08817a070","Type":"ContainerStarted","Data":"0d2a9d6a1537ed6afb6b28b706ec6bd7a7cff3f280a01827c713167abdd234d6"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.678044 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" event={"ID":"dc5fcae4-29ef-4706-9b51-e4f3e99f1c36","Type":"ContainerStarted","Data":"212fa8311de9d3b8c3a25af06b2aa731a51c713c0f4ec62061fda456170d1c8e"} Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.678410 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg" podUID="db247d6e-0199-4fde-85d9-86b08817a070" Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.679081 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" podUID="dc5fcae4-29ef-4706-9b51-e4f3e99f1c36" Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.679252 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6" event={"ID":"cbc0c461-56e5-4386-bb50-a633b89911e0","Type":"ContainerStarted","Data":"f0eba6e7ebf39ae7487caeffa300ba6fec00d190aa65bce9110617724944c383"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.680119 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh" event={"ID":"1bee3920-4de9-436c-8d80-4ea7f0903f11","Type":"ContainerStarted","Data":"f068b58c9bc6a1a60301b8d117bc65071670c23286ed2ed5b748c2e665e15d1e"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.695779 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5" event={"ID":"99b546dc-24de-4e98-9fce-7d457bf981ec","Type":"ContainerStarted","Data":"90d4dfd8d3f901dc54825b3888d65aa6bd8dc0b70df7253fd58a40fa787d6c2c"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.699658 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz" event={"ID":"56f5f54a-9f54-4e6e-a9fc-129398f0df0b","Type":"ContainerStarted","Data":"0e8e24b5c9b34bba9ceafb78988b837f137f409dac68a2d45df3feb6075184e1"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.707932 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl" event={"ID":"60510787-a5c0-467a-969f-38c6be9b2184","Type":"ContainerStarted","Data":"12149d0d672fe1aae2406409b1fbd70287ed5f56f33c2eed3d10f13957d07117"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.737079 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84" event={"ID":"00926e2b-9085-412d-9b65-2c71f76bd9ed","Type":"ContainerStarted","Data":"4f2f98ed637845cf906d01d7ecce9c1fba4aa8e4a7a6e640acb383c38328d1f0"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.743129 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" event={"ID":"095020a1-2516-42f1-9223-339fd943c52f","Type":"ContainerStarted","Data":"39bc0df625306498312f3879ac4b28fecd8b3a5ff303082dcb721c5869298968"} Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.746260 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" podUID="095020a1-2516-42f1-9223-339fd943c52f" Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.746906 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm" event={"ID":"cae81ff3-9955-429b-9e24-3093d3c7f1fa","Type":"ContainerStarted","Data":"7e3e07022e4cdf3f98ad9b72a0d5d30646f4cb7bec59798afeebd86ea8663331"} Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.919994 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:39 crc kubenswrapper[4968]: I0218 15:37:39.920164 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.920296 4968 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.920488 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs podName:5e48a888-178c-4d24-b960-60b6251940db nodeName:}" failed. No retries permitted until 2026-02-18 15:37:41.920466334 +0000 UTC m=+961.305911206 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs") pod "openstack-operator-controller-manager-79f9cfbcf7-bzx4g" (UID: "5e48a888-178c-4d24-b960-60b6251940db") : secret "metrics-server-cert" not found Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.920573 4968 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 18 15:37:39 crc kubenswrapper[4968]: E0218 15:37:39.921015 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs podName:5e48a888-178c-4d24-b960-60b6251940db nodeName:}" failed. No retries permitted until 2026-02-18 15:37:41.920999829 +0000 UTC m=+961.306444711 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs") pod "openstack-operator-controller-manager-79f9cfbcf7-bzx4g" (UID: "5e48a888-178c-4d24-b960-60b6251940db") : secret "webhook-server-cert" not found Feb 18 15:37:40 crc kubenswrapper[4968]: E0218 15:37:40.755836 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" podUID="53297588-6688-444d-952c-810b9b265acf" Feb 18 15:37:40 crc kubenswrapper[4968]: E0218 15:37:40.756233 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg" podUID="db247d6e-0199-4fde-85d9-86b08817a070" Feb 18 15:37:40 crc kubenswrapper[4968]: E0218 15:37:40.756783 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" podUID="964d8817-99e3-4a8d-91ab-876ec187f62d" Feb 18 15:37:40 crc kubenswrapper[4968]: E0218 15:37:40.757220 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" podUID="095020a1-2516-42f1-9223-339fd943c52f" Feb 18 15:37:40 crc kubenswrapper[4968]: E0218 15:37:40.757361 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" podUID="e39cd00d-1815-442a-9343-277280cf8ccc" Feb 18 15:37:40 crc kubenswrapper[4968]: E0218 15:37:40.758726 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" podUID="dc5fcae4-29ef-4706-9b51-e4f3e99f1c36" Feb 18 15:37:41 crc kubenswrapper[4968]: I0218 15:37:41.251199 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert\") pod \"infra-operator-controller-manager-79d975b745-lwn4x\" (UID: \"fb1b11dc-0a5b-4784-bc3b-a0976c481831\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:37:41 crc kubenswrapper[4968]: E0218 15:37:41.251394 4968 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 18 15:37:41 crc kubenswrapper[4968]: E0218 15:37:41.251621 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert podName:fb1b11dc-0a5b-4784-bc3b-a0976c481831 nodeName:}" failed. No retries permitted until 2026-02-18 15:37:45.251602384 +0000 UTC m=+964.637047246 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert") pod "infra-operator-controller-manager-79d975b745-lwn4x" (UID: "fb1b11dc-0a5b-4784-bc3b-a0976c481831") : secret "infra-operator-webhook-server-cert" not found Feb 18 15:37:41 crc kubenswrapper[4968]: I0218 15:37:41.556170 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz\" (UID: \"c2c93869-c328-4b28-8098-18191433d4d2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:37:41 crc kubenswrapper[4968]: E0218 15:37:41.556313 4968 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 15:37:41 crc kubenswrapper[4968]: E0218 15:37:41.556355 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert podName:c2c93869-c328-4b28-8098-18191433d4d2 nodeName:}" failed. No retries permitted until 2026-02-18 15:37:45.55634296 +0000 UTC m=+964.941787822 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" (UID: "c2c93869-c328-4b28-8098-18191433d4d2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 15:37:41 crc kubenswrapper[4968]: I0218 15:37:41.960035 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:41 crc kubenswrapper[4968]: I0218 15:37:41.960150 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:41 crc kubenswrapper[4968]: E0218 15:37:41.960256 4968 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 18 15:37:41 crc kubenswrapper[4968]: E0218 15:37:41.960289 4968 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 18 15:37:41 crc kubenswrapper[4968]: E0218 15:37:41.960353 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs podName:5e48a888-178c-4d24-b960-60b6251940db nodeName:}" failed. No retries permitted until 2026-02-18 15:37:45.960331142 +0000 UTC m=+965.345776074 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs") pod "openstack-operator-controller-manager-79f9cfbcf7-bzx4g" (UID: "5e48a888-178c-4d24-b960-60b6251940db") : secret "metrics-server-cert" not found Feb 18 15:37:41 crc kubenswrapper[4968]: E0218 15:37:41.960421 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs podName:5e48a888-178c-4d24-b960-60b6251940db nodeName:}" failed. No retries permitted until 2026-02-18 15:37:45.960377513 +0000 UTC m=+965.345822375 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs") pod "openstack-operator-controller-manager-79f9cfbcf7-bzx4g" (UID: "5e48a888-178c-4d24-b960-60b6251940db") : secret "webhook-server-cert" not found Feb 18 15:37:44 crc kubenswrapper[4968]: I0218 15:37:44.367314 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:37:44 crc kubenswrapper[4968]: I0218 15:37:44.367369 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:37:44 crc kubenswrapper[4968]: I0218 15:37:44.367409 4968 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:37:44 crc kubenswrapper[4968]: I0218 15:37:44.367909 4968 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2024fbd9a3a558a0c7ebda43f30fc1d9386f6d00027774f8f91dd895439fc899"} pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:37:44 crc kubenswrapper[4968]: I0218 15:37:44.367952 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" containerID="cri-o://2024fbd9a3a558a0c7ebda43f30fc1d9386f6d00027774f8f91dd895439fc899" gracePeriod=600 Feb 18 15:37:45 crc kubenswrapper[4968]: I0218 15:37:45.311094 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert\") pod \"infra-operator-controller-manager-79d975b745-lwn4x\" (UID: \"fb1b11dc-0a5b-4784-bc3b-a0976c481831\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:37:45 crc kubenswrapper[4968]: E0218 15:37:45.311265 4968 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 18 15:37:45 crc kubenswrapper[4968]: E0218 15:37:45.311328 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert podName:fb1b11dc-0a5b-4784-bc3b-a0976c481831 nodeName:}" failed. No retries permitted until 2026-02-18 15:37:53.311312329 +0000 UTC m=+972.696757191 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert") pod "infra-operator-controller-manager-79d975b745-lwn4x" (UID: "fb1b11dc-0a5b-4784-bc3b-a0976c481831") : secret "infra-operator-webhook-server-cert" not found Feb 18 15:37:45 crc kubenswrapper[4968]: I0218 15:37:45.615045 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz\" (UID: \"c2c93869-c328-4b28-8098-18191433d4d2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:37:45 crc kubenswrapper[4968]: E0218 15:37:45.615210 4968 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 15:37:45 crc kubenswrapper[4968]: E0218 15:37:45.615445 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert podName:c2c93869-c328-4b28-8098-18191433d4d2 nodeName:}" failed. No retries permitted until 2026-02-18 15:37:53.615425816 +0000 UTC m=+973.000870678 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" (UID: "c2c93869-c328-4b28-8098-18191433d4d2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 15:37:45 crc kubenswrapper[4968]: I0218 15:37:45.797917 4968 generic.go:334] "Generic (PLEG): container finished" podID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerID="2024fbd9a3a558a0c7ebda43f30fc1d9386f6d00027774f8f91dd895439fc899" exitCode=0 Feb 18 15:37:45 crc kubenswrapper[4968]: I0218 15:37:45.797977 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerDied","Data":"2024fbd9a3a558a0c7ebda43f30fc1d9386f6d00027774f8f91dd895439fc899"} Feb 18 15:37:45 crc kubenswrapper[4968]: I0218 15:37:45.798051 4968 scope.go:117] "RemoveContainer" containerID="47c8c50ad8111d46dcdc1997d2e8171ff7419555273a45fc7043517f33afc7b7" Feb 18 15:37:46 crc kubenswrapper[4968]: I0218 15:37:46.022353 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:46 crc kubenswrapper[4968]: I0218 15:37:46.022439 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:46 crc kubenswrapper[4968]: E0218 15:37:46.022558 4968 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 18 15:37:46 crc kubenswrapper[4968]: E0218 15:37:46.022564 4968 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 18 15:37:46 crc kubenswrapper[4968]: E0218 15:37:46.022606 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs podName:5e48a888-178c-4d24-b960-60b6251940db nodeName:}" failed. No retries permitted until 2026-02-18 15:37:54.022592547 +0000 UTC m=+973.408037409 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs") pod "openstack-operator-controller-manager-79f9cfbcf7-bzx4g" (UID: "5e48a888-178c-4d24-b960-60b6251940db") : secret "metrics-server-cert" not found Feb 18 15:37:46 crc kubenswrapper[4968]: E0218 15:37:46.022715 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs podName:5e48a888-178c-4d24-b960-60b6251940db nodeName:}" failed. No retries permitted until 2026-02-18 15:37:54.022690959 +0000 UTC m=+973.408135871 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs") pod "openstack-operator-controller-manager-79f9cfbcf7-bzx4g" (UID: "5e48a888-178c-4d24-b960-60b6251940db") : secret "webhook-server-cert" not found Feb 18 15:37:53 crc kubenswrapper[4968]: I0218 15:37:53.360343 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert\") pod \"infra-operator-controller-manager-79d975b745-lwn4x\" (UID: \"fb1b11dc-0a5b-4784-bc3b-a0976c481831\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:37:53 crc kubenswrapper[4968]: E0218 15:37:53.360561 4968 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 18 15:37:53 crc kubenswrapper[4968]: E0218 15:37:53.361014 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert podName:fb1b11dc-0a5b-4784-bc3b-a0976c481831 nodeName:}" failed. No retries permitted until 2026-02-18 15:38:09.360996945 +0000 UTC m=+988.746441807 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert") pod "infra-operator-controller-manager-79d975b745-lwn4x" (UID: "fb1b11dc-0a5b-4784-bc3b-a0976c481831") : secret "infra-operator-webhook-server-cert" not found Feb 18 15:37:53 crc kubenswrapper[4968]: I0218 15:37:53.664428 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz\" (UID: \"c2c93869-c328-4b28-8098-18191433d4d2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:37:53 crc kubenswrapper[4968]: E0218 15:37:53.664663 4968 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 15:37:53 crc kubenswrapper[4968]: E0218 15:37:53.664789 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert podName:c2c93869-c328-4b28-8098-18191433d4d2 nodeName:}" failed. No retries permitted until 2026-02-18 15:38:09.664741002 +0000 UTC m=+989.050185884 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" (UID: "c2c93869-c328-4b28-8098-18191433d4d2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 18 15:37:54 crc kubenswrapper[4968]: I0218 15:37:54.070552 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:54 crc kubenswrapper[4968]: I0218 15:37:54.070691 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:54 crc kubenswrapper[4968]: E0218 15:37:54.070871 4968 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 18 15:37:54 crc kubenswrapper[4968]: E0218 15:37:54.070919 4968 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs podName:5e48a888-178c-4d24-b960-60b6251940db nodeName:}" failed. No retries permitted until 2026-02-18 15:38:10.070905544 +0000 UTC m=+989.456350406 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs") pod "openstack-operator-controller-manager-79f9cfbcf7-bzx4g" (UID: "5e48a888-178c-4d24-b960-60b6251940db") : secret "metrics-server-cert" not found Feb 18 15:37:54 crc kubenswrapper[4968]: I0218 15:37:54.076523 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-webhook-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:37:55 crc kubenswrapper[4968]: I0218 15:37:55.887461 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerStarted","Data":"1c0ee8bcfd29be60e6cb05f3e8eb4433c41144c8013553fdaf58be5bb39a80b9"} Feb 18 15:37:55 crc kubenswrapper[4968]: I0218 15:37:55.888723 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5" event={"ID":"99b546dc-24de-4e98-9fce-7d457bf981ec","Type":"ContainerStarted","Data":"b89312a1a1faf4d42e92f5b57f665151484f4e8c0d666fba411e395eb6b6a7d7"} Feb 18 15:37:55 crc kubenswrapper[4968]: I0218 15:37:55.889264 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5" Feb 18 15:37:55 crc kubenswrapper[4968]: I0218 15:37:55.896106 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62" event={"ID":"35c5d058-dc84-4faf-a645-c8463d0e6eb1","Type":"ContainerStarted","Data":"edb1e01bd40f9e67115e72e42f35e0b8b62758e010debbf07109fb55c24fb552"} Feb 18 15:37:55 crc kubenswrapper[4968]: I0218 15:37:55.896770 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62" Feb 18 15:37:55 crc kubenswrapper[4968]: I0218 15:37:55.925216 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5" podStartSLOduration=3.788667824 podStartE2EDuration="18.92519737s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:38.96055113 +0000 UTC m=+958.345995992" lastFinishedPulling="2026-02-18 15:37:54.097080676 +0000 UTC m=+973.482525538" observedRunningTime="2026-02-18 15:37:55.919566323 +0000 UTC m=+975.305011185" watchObservedRunningTime="2026-02-18 15:37:55.92519737 +0000 UTC m=+975.310642242" Feb 18 15:37:55 crc kubenswrapper[4968]: I0218 15:37:55.949315 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62" podStartSLOduration=3.956462158 podStartE2EDuration="18.949300625s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:38.511685033 +0000 UTC m=+957.897129895" lastFinishedPulling="2026-02-18 15:37:53.50452349 +0000 UTC m=+972.889968362" observedRunningTime="2026-02-18 15:37:55.942630688 +0000 UTC m=+975.328075540" watchObservedRunningTime="2026-02-18 15:37:55.949300625 +0000 UTC m=+975.334745487" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.909277 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" event={"ID":"095020a1-2516-42f1-9223-339fd943c52f","Type":"ContainerStarted","Data":"bcdcc87f2d68ad5ded2a3565df386ec2531694868b35daffa80fb5f4525d5fff"} Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.909949 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.911505 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf" event={"ID":"debbb308-3378-47b6-8e2a-b0d8edc95bfc","Type":"ContainerStarted","Data":"10d383e419985829f7620671b6a733d0f430fb39fe36a808011ad0a234e93c84"} Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.911879 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.916788 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t" event={"ID":"8a5aaad5-3e01-4de2-9bf1-e64680ce7b18","Type":"ContainerStarted","Data":"64632c046c44d313690605dfb9526c39086ba0afda2132bc7cb280c986e9096c"} Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.916938 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.921788 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84" event={"ID":"00926e2b-9085-412d-9b65-2c71f76bd9ed","Type":"ContainerStarted","Data":"c6500d31c5b7eb3a1d3dc17570453f432e5594adf483705143b9e5cedf6ad5ac"} Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.921898 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.940886 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj" event={"ID":"6f7c68fd-afbb-4751-b1da-d27bfe32d6c7","Type":"ContainerStarted","Data":"7b5ef9f3986e0cb5c7c08ab99dfd885cf300f39f882fbf7e58fc36a157befad2"} Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.941489 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.945468 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6" event={"ID":"cbc0c461-56e5-4386-bb50-a633b89911e0","Type":"ContainerStarted","Data":"5bf8d6cd2b6f3a81a4b67ff953ba04304d069ad2eb9ea3de50267760a3a86edf"} Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.946361 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.953550 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh" event={"ID":"1bee3920-4de9-436c-8d80-4ea7f0903f11","Type":"ContainerStarted","Data":"ce6bf6532e04195be44940f0b7f1ded905892de84904844b279ab2a9408f299f"} Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.954085 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.955290 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl" event={"ID":"60510787-a5c0-467a-969f-38c6be9b2184","Type":"ContainerStarted","Data":"5a9eb967553a2e62d7178dea285590c6a11ba2579f95d5315523f57b4e5b9466"} Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.955663 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.960097 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj" event={"ID":"aa54cd8f-435a-4055-9668-1ab7c3e72516","Type":"ContainerStarted","Data":"33ce10f57f2a4bf94cc09b99cf946e3657ff1cbcd0f64c0057ba99b882f201b8"} Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.960191 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.962008 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v" event={"ID":"87fee87a-b2aa-46b7-8f11-89b00fe7638c","Type":"ContainerStarted","Data":"20d4c1d28a57d74c03ba62bff14e092718d657b3064dde819f4574428a3de997"} Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.962341 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.973574 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" podStartSLOduration=3.607155626 podStartE2EDuration="19.973556149s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.487495122 +0000 UTC m=+958.872939984" lastFinishedPulling="2026-02-18 15:37:55.853895645 +0000 UTC m=+975.239340507" observedRunningTime="2026-02-18 15:37:56.941207004 +0000 UTC m=+976.326651876" watchObservedRunningTime="2026-02-18 15:37:56.973556149 +0000 UTC m=+976.359001011" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.973977 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" event={"ID":"dc5fcae4-29ef-4706-9b51-e4f3e99f1c36","Type":"ContainerStarted","Data":"6d6a16e14ae47dade20c45203f2bc118c3a265318a3392cf0d113855b0958ddc"} Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.974688 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.977168 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf" podStartSLOduration=4.210517146 podStartE2EDuration="19.97714893s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.045330632 +0000 UTC m=+958.430775494" lastFinishedPulling="2026-02-18 15:37:54.811962406 +0000 UTC m=+974.197407278" observedRunningTime="2026-02-18 15:37:56.973959991 +0000 UTC m=+976.359404853" watchObservedRunningTime="2026-02-18 15:37:56.97714893 +0000 UTC m=+976.362593792" Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.991353 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz" event={"ID":"56f5f54a-9f54-4e6e-a9fc-129398f0df0b","Type":"ContainerStarted","Data":"686bac63dfab0c1601b374d3fd2618bbfd612cbab14b8d6657249ba9fa0661c5"} Feb 18 15:37:56 crc kubenswrapper[4968]: I0218 15:37:56.992022 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.003293 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" event={"ID":"e39cd00d-1815-442a-9343-277280cf8ccc","Type":"ContainerStarted","Data":"8c074b98c962bcf7f25848140e3126fee8513c6ae0c4d31695a8b95b6c19d07e"} Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.004359 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.012366 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj" podStartSLOduration=4.977216084 podStartE2EDuration="20.012344814s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.062494942 +0000 UTC m=+958.447939804" lastFinishedPulling="2026-02-18 15:37:54.097623672 +0000 UTC m=+973.483068534" observedRunningTime="2026-02-18 15:37:57.011762468 +0000 UTC m=+976.397207330" watchObservedRunningTime="2026-02-18 15:37:57.012344814 +0000 UTC m=+976.397789676" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.013670 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm" event={"ID":"cae81ff3-9955-429b-9e24-3093d3c7f1fa","Type":"ContainerStarted","Data":"6822c865ff8d99d7f7bf0e6c0ce3bec5420173268d124651bb7d1a543935d56e"} Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.014306 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.036736 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84" podStartSLOduration=4.280503773 podStartE2EDuration="20.036719546s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.056855124 +0000 UTC m=+958.442299986" lastFinishedPulling="2026-02-18 15:37:54.813070897 +0000 UTC m=+974.198515759" observedRunningTime="2026-02-18 15:37:57.031361856 +0000 UTC m=+976.416806718" watchObservedRunningTime="2026-02-18 15:37:57.036719546 +0000 UTC m=+976.422164408" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.039333 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c" event={"ID":"392e8c65-7168-4e8b-b672-fb0b53da6ad5","Type":"ContainerStarted","Data":"d7afd6b1d07a86c3ac5973014d07e2e2ef5079fdb6d8bf9f65a4d5eed9942b0a"} Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.039370 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.053407 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t" podStartSLOduration=5.383926713 podStartE2EDuration="20.053393163s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.42737155 +0000 UTC m=+958.812816412" lastFinishedPulling="2026-02-18 15:37:54.096838 +0000 UTC m=+973.482282862" observedRunningTime="2026-02-18 15:37:57.050095571 +0000 UTC m=+976.435540433" watchObservedRunningTime="2026-02-18 15:37:57.053393163 +0000 UTC m=+976.438838015" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.086886 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v" podStartSLOduration=4.9611309850000005 podStartE2EDuration="20.086733296s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:38.99951055 +0000 UTC m=+958.384955412" lastFinishedPulling="2026-02-18 15:37:54.125112861 +0000 UTC m=+973.510557723" observedRunningTime="2026-02-18 15:37:57.082286841 +0000 UTC m=+976.467731713" watchObservedRunningTime="2026-02-18 15:37:57.086733296 +0000 UTC m=+976.472178158" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.112865 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj" podStartSLOduration=4.720713869 podStartE2EDuration="20.112849956s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.423022978 +0000 UTC m=+958.808467840" lastFinishedPulling="2026-02-18 15:37:54.815159065 +0000 UTC m=+974.200603927" observedRunningTime="2026-02-18 15:37:57.108118824 +0000 UTC m=+976.493563696" watchObservedRunningTime="2026-02-18 15:37:57.112849956 +0000 UTC m=+976.498294818" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.139547 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6" podStartSLOduration=4.352622181 podStartE2EDuration="20.139526322s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.030036164 +0000 UTC m=+958.415481026" lastFinishedPulling="2026-02-18 15:37:54.816940305 +0000 UTC m=+974.202385167" observedRunningTime="2026-02-18 15:37:57.135223732 +0000 UTC m=+976.520668594" watchObservedRunningTime="2026-02-18 15:37:57.139526322 +0000 UTC m=+976.524971184" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.161268 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" podStartSLOduration=4.741071198 podStartE2EDuration="20.16125169s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.49922921 +0000 UTC m=+958.884674072" lastFinishedPulling="2026-02-18 15:37:54.919409702 +0000 UTC m=+974.304854564" observedRunningTime="2026-02-18 15:37:57.156153768 +0000 UTC m=+976.541598630" watchObservedRunningTime="2026-02-18 15:37:57.16125169 +0000 UTC m=+976.546696552" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.184973 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh" podStartSLOduration=6.097512646 podStartE2EDuration="20.184961004s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.457476832 +0000 UTC m=+958.842921694" lastFinishedPulling="2026-02-18 15:37:53.54492519 +0000 UTC m=+972.930370052" observedRunningTime="2026-02-18 15:37:57.183951675 +0000 UTC m=+976.569396537" watchObservedRunningTime="2026-02-18 15:37:57.184961004 +0000 UTC m=+976.570405866" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.222117 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c" podStartSLOduration=3.918161727 podStartE2EDuration="20.222090972s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:38.511190489 +0000 UTC m=+957.896635351" lastFinishedPulling="2026-02-18 15:37:54.815119734 +0000 UTC m=+974.200564596" observedRunningTime="2026-02-18 15:37:57.212158795 +0000 UTC m=+976.597603647" watchObservedRunningTime="2026-02-18 15:37:57.222090972 +0000 UTC m=+976.607535834" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.239436 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" podStartSLOduration=3.849329811 podStartE2EDuration="20.239415417s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.478438909 +0000 UTC m=+958.863883771" lastFinishedPulling="2026-02-18 15:37:55.868524515 +0000 UTC m=+975.253969377" observedRunningTime="2026-02-18 15:37:57.23309711 +0000 UTC m=+976.618541972" watchObservedRunningTime="2026-02-18 15:37:57.239415417 +0000 UTC m=+976.624860279" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.281297 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm" podStartSLOduration=4.480581441 podStartE2EDuration="20.281276878s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.02560083 +0000 UTC m=+958.411045692" lastFinishedPulling="2026-02-18 15:37:54.826296267 +0000 UTC m=+974.211741129" observedRunningTime="2026-02-18 15:37:57.275702892 +0000 UTC m=+976.661147754" watchObservedRunningTime="2026-02-18 15:37:57.281276878 +0000 UTC m=+976.666721740" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.317435 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl" podStartSLOduration=4.497832503 podStartE2EDuration="20.317419059s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.007407181 +0000 UTC m=+958.392852033" lastFinishedPulling="2026-02-18 15:37:54.826993727 +0000 UTC m=+974.212438589" observedRunningTime="2026-02-18 15:37:57.31707842 +0000 UTC m=+976.702523282" watchObservedRunningTime="2026-02-18 15:37:57.317419059 +0000 UTC m=+976.702863921" Feb 18 15:37:57 crc kubenswrapper[4968]: I0218 15:37:57.347044 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz" podStartSLOduration=6.298612272 podStartE2EDuration="20.347029098s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.456085553 +0000 UTC m=+958.841530415" lastFinishedPulling="2026-02-18 15:37:53.504502379 +0000 UTC m=+972.889947241" observedRunningTime="2026-02-18 15:37:57.3456722 +0000 UTC m=+976.731117062" watchObservedRunningTime="2026-02-18 15:37:57.347029098 +0000 UTC m=+976.732473960" Feb 18 15:38:03 crc kubenswrapper[4968]: I0218 15:38:03.091294 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" event={"ID":"964d8817-99e3-4a8d-91ab-876ec187f62d","Type":"ContainerStarted","Data":"07d616c0005f5b2bc1f9795ae0853823c49e52adb06b380293cd1ec66e0daa45"} Feb 18 15:38:03 crc kubenswrapper[4968]: I0218 15:38:03.092098 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" Feb 18 15:38:03 crc kubenswrapper[4968]: I0218 15:38:03.092942 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg" event={"ID":"db247d6e-0199-4fde-85d9-86b08817a070","Type":"ContainerStarted","Data":"959e607e767e4a87a4526675c9f5eff614263351372ff12febbde512d35d01d1"} Feb 18 15:38:03 crc kubenswrapper[4968]: I0218 15:38:03.094236 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" event={"ID":"53297588-6688-444d-952c-810b9b265acf","Type":"ContainerStarted","Data":"8cf699e95e90d795d394e9a14eb6000e2e9af5282ba4148694399d3cf0435d35"} Feb 18 15:38:03 crc kubenswrapper[4968]: I0218 15:38:03.094441 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" Feb 18 15:38:03 crc kubenswrapper[4968]: I0218 15:38:03.114663 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" podStartSLOduration=3.524540434 podStartE2EDuration="26.114644822s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.46634872 +0000 UTC m=+958.851793582" lastFinishedPulling="2026-02-18 15:38:02.056453108 +0000 UTC m=+981.441897970" observedRunningTime="2026-02-18 15:38:03.11279007 +0000 UTC m=+982.498234942" watchObservedRunningTime="2026-02-18 15:38:03.114644822 +0000 UTC m=+982.500089684" Feb 18 15:38:03 crc kubenswrapper[4968]: I0218 15:38:03.135030 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-98jxg" podStartSLOduration=2.516461249 podStartE2EDuration="25.135006412s" podCreationTimestamp="2026-02-18 15:37:38 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.498388557 +0000 UTC m=+958.883833419" lastFinishedPulling="2026-02-18 15:38:02.11693369 +0000 UTC m=+981.502378582" observedRunningTime="2026-02-18 15:38:03.128100708 +0000 UTC m=+982.513545560" watchObservedRunningTime="2026-02-18 15:38:03.135006412 +0000 UTC m=+982.520451274" Feb 18 15:38:07 crc kubenswrapper[4968]: I0218 15:38:07.552018 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-7rw4c" Feb 18 15:38:07 crc kubenswrapper[4968]: I0218 15:38:07.582825 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-8ng62" Feb 18 15:38:07 crc kubenswrapper[4968]: I0218 15:38:07.594867 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" podStartSLOduration=7.977978944 podStartE2EDuration="30.594838729s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:37:39.471641979 +0000 UTC m=+958.857086841" lastFinishedPulling="2026-02-18 15:38:02.088501764 +0000 UTC m=+981.473946626" observedRunningTime="2026-02-18 15:38:03.146526054 +0000 UTC m=+982.531970916" watchObservedRunningTime="2026-02-18 15:38:07.594838729 +0000 UTC m=+986.980283641" Feb 18 15:38:07 crc kubenswrapper[4968]: I0218 15:38:07.645199 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-zc86v" Feb 18 15:38:07 crc kubenswrapper[4968]: I0218 15:38:07.682323 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987464f4-7ltz5" Feb 18 15:38:07 crc kubenswrapper[4968]: I0218 15:38:07.719815 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-kkzx6" Feb 18 15:38:07 crc kubenswrapper[4968]: I0218 15:38:07.739508 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-v9gsm" Feb 18 15:38:07 crc kubenswrapper[4968]: I0218 15:38:07.792910 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-6jkdl" Feb 18 15:38:07 crc kubenswrapper[4968]: I0218 15:38:07.807562 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-56xzj" Feb 18 15:38:07 crc kubenswrapper[4968]: I0218 15:38:07.939020 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-f4prf" Feb 18 15:38:07 crc kubenswrapper[4968]: I0218 15:38:07.970674 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-mbn84" Feb 18 15:38:08 crc kubenswrapper[4968]: I0218 15:38:08.022288 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vvqqh" Feb 18 15:38:08 crc kubenswrapper[4968]: I0218 15:38:08.053447 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-xhwzj" Feb 18 15:38:08 crc kubenswrapper[4968]: I0218 15:38:08.150819 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-jhq7j" Feb 18 15:38:08 crc kubenswrapper[4968]: I0218 15:38:08.303631 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-wbp4k" Feb 18 15:38:08 crc kubenswrapper[4968]: I0218 15:38:08.318554 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-kskjh" Feb 18 15:38:08 crc kubenswrapper[4968]: I0218 15:38:08.328623 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68f46476f-n6bn4" Feb 18 15:38:08 crc kubenswrapper[4968]: I0218 15:38:08.366903 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-dfg2t" Feb 18 15:38:08 crc kubenswrapper[4968]: I0218 15:38:08.397878 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7866795846-qphmd" Feb 18 15:38:08 crc kubenswrapper[4968]: I0218 15:38:08.463726 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5c48998cbf-jdwxz" Feb 18 15:38:09 crc kubenswrapper[4968]: I0218 15:38:09.422597 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert\") pod \"infra-operator-controller-manager-79d975b745-lwn4x\" (UID: \"fb1b11dc-0a5b-4784-bc3b-a0976c481831\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:38:09 crc kubenswrapper[4968]: I0218 15:38:09.435735 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb1b11dc-0a5b-4784-bc3b-a0976c481831-cert\") pod \"infra-operator-controller-manager-79d975b745-lwn4x\" (UID: \"fb1b11dc-0a5b-4784-bc3b-a0976c481831\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:38:09 crc kubenswrapper[4968]: I0218 15:38:09.555883 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-92p4n" Feb 18 15:38:09 crc kubenswrapper[4968]: I0218 15:38:09.564286 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:38:09 crc kubenswrapper[4968]: I0218 15:38:09.725041 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz\" (UID: \"c2c93869-c328-4b28-8098-18191433d4d2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:38:09 crc kubenswrapper[4968]: I0218 15:38:09.733580 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2c93869-c328-4b28-8098-18191433d4d2-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz\" (UID: \"c2c93869-c328-4b28-8098-18191433d4d2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:38:10 crc kubenswrapper[4968]: I0218 15:38:10.001422 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-dmhbs" Feb 18 15:38:10 crc kubenswrapper[4968]: I0218 15:38:10.009156 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:38:10 crc kubenswrapper[4968]: I0218 15:38:10.070843 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x"] Feb 18 15:38:10 crc kubenswrapper[4968]: W0218 15:38:10.074643 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb1b11dc_0a5b_4784_bc3b_a0976c481831.slice/crio-6a39e04fd38bd839bc9133d93522ef4a13aba3d517c46a487d427ba351c9c815 WatchSource:0}: Error finding container 6a39e04fd38bd839bc9133d93522ef4a13aba3d517c46a487d427ba351c9c815: Status 404 returned error can't find the container with id 6a39e04fd38bd839bc9133d93522ef4a13aba3d517c46a487d427ba351c9c815 Feb 18 15:38:10 crc kubenswrapper[4968]: I0218 15:38:10.129505 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:38:10 crc kubenswrapper[4968]: I0218 15:38:10.135127 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e48a888-178c-4d24-b960-60b6251940db-metrics-certs\") pod \"openstack-operator-controller-manager-79f9cfbcf7-bzx4g\" (UID: \"5e48a888-178c-4d24-b960-60b6251940db\") " pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:38:10 crc kubenswrapper[4968]: I0218 15:38:10.148666 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" event={"ID":"fb1b11dc-0a5b-4784-bc3b-a0976c481831","Type":"ContainerStarted","Data":"6a39e04fd38bd839bc9133d93522ef4a13aba3d517c46a487d427ba351c9c815"} Feb 18 15:38:10 crc kubenswrapper[4968]: I0218 15:38:10.275423 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-r462p" Feb 18 15:38:10 crc kubenswrapper[4968]: I0218 15:38:10.283694 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:38:10 crc kubenswrapper[4968]: I0218 15:38:10.306111 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz"] Feb 18 15:38:10 crc kubenswrapper[4968]: W0218 15:38:10.317222 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2c93869_c328_4b28_8098_18191433d4d2.slice/crio-5bec4b1bbbd10105a0da129d061d0d73c09341dcd7d0655be615ac457e137928 WatchSource:0}: Error finding container 5bec4b1bbbd10105a0da129d061d0d73c09341dcd7d0655be615ac457e137928: Status 404 returned error can't find the container with id 5bec4b1bbbd10105a0da129d061d0d73c09341dcd7d0655be615ac457e137928 Feb 18 15:38:10 crc kubenswrapper[4968]: I0218 15:38:10.592949 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g"] Feb 18 15:38:10 crc kubenswrapper[4968]: W0218 15:38:10.599573 4968 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e48a888_178c_4d24_b960_60b6251940db.slice/crio-010e24a74e0c179e9f5fe467ec77848448e5ac25e116937aaffe7659d7e6443d WatchSource:0}: Error finding container 010e24a74e0c179e9f5fe467ec77848448e5ac25e116937aaffe7659d7e6443d: Status 404 returned error can't find the container with id 010e24a74e0c179e9f5fe467ec77848448e5ac25e116937aaffe7659d7e6443d Feb 18 15:38:11 crc kubenswrapper[4968]: I0218 15:38:11.166788 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" event={"ID":"c2c93869-c328-4b28-8098-18191433d4d2","Type":"ContainerStarted","Data":"5bec4b1bbbd10105a0da129d061d0d73c09341dcd7d0655be615ac457e137928"} Feb 18 15:38:11 crc kubenswrapper[4968]: I0218 15:38:11.169183 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" event={"ID":"5e48a888-178c-4d24-b960-60b6251940db","Type":"ContainerStarted","Data":"010e24a74e0c179e9f5fe467ec77848448e5ac25e116937aaffe7659d7e6443d"} Feb 18 15:38:16 crc kubenswrapper[4968]: I0218 15:38:16.212684 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" event={"ID":"5e48a888-178c-4d24-b960-60b6251940db","Type":"ContainerStarted","Data":"201eab67aa973375bce214730a97bd4e0b28f582179bb6998b78b5d8f371a310"} Feb 18 15:38:16 crc kubenswrapper[4968]: I0218 15:38:16.213607 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:38:16 crc kubenswrapper[4968]: I0218 15:38:16.259330 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" podStartSLOduration=39.259293486 podStartE2EDuration="39.259293486s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:38:16.246252231 +0000 UTC m=+995.631697113" watchObservedRunningTime="2026-02-18 15:38:16.259293486 +0000 UTC m=+995.644738388" Feb 18 15:38:18 crc kubenswrapper[4968]: I0218 15:38:18.230777 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" event={"ID":"c2c93869-c328-4b28-8098-18191433d4d2","Type":"ContainerStarted","Data":"21519996948d76e83cfb9cc7544417df813b591481c315bd6b3873917ffdb937"} Feb 18 15:38:18 crc kubenswrapper[4968]: I0218 15:38:18.232360 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:38:19 crc kubenswrapper[4968]: I0218 15:38:19.240079 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" event={"ID":"fb1b11dc-0a5b-4784-bc3b-a0976c481831","Type":"ContainerStarted","Data":"0f811ea29d07fb47359693cafe82b6663a6be9c75b5b66005862b68a8d3abb52"} Feb 18 15:38:19 crc kubenswrapper[4968]: I0218 15:38:19.266556 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" podStartSLOduration=34.940063961999996 podStartE2EDuration="42.266535957s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:38:10.324228887 +0000 UTC m=+989.709673759" lastFinishedPulling="2026-02-18 15:38:17.650700892 +0000 UTC m=+997.036145754" observedRunningTime="2026-02-18 15:38:18.266220052 +0000 UTC m=+997.651664924" watchObservedRunningTime="2026-02-18 15:38:19.266535957 +0000 UTC m=+998.651980819" Feb 18 15:38:19 crc kubenswrapper[4968]: I0218 15:38:19.267995 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" podStartSLOduration=33.876531927 podStartE2EDuration="42.267984097s" podCreationTimestamp="2026-02-18 15:37:37 +0000 UTC" firstStartedPulling="2026-02-18 15:38:10.078334867 +0000 UTC m=+989.463779739" lastFinishedPulling="2026-02-18 15:38:18.469787047 +0000 UTC m=+997.855231909" observedRunningTime="2026-02-18 15:38:19.260091296 +0000 UTC m=+998.645536168" watchObservedRunningTime="2026-02-18 15:38:19.267984097 +0000 UTC m=+998.653428959" Feb 18 15:38:19 crc kubenswrapper[4968]: I0218 15:38:19.565074 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:38:29 crc kubenswrapper[4968]: I0218 15:38:29.572766 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79d975b745-lwn4x" Feb 18 15:38:30 crc kubenswrapper[4968]: I0218 15:38:30.017591 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz" Feb 18 15:38:30 crc kubenswrapper[4968]: I0218 15:38:30.293367 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-79f9cfbcf7-bzx4g" Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.094506 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ftfvg/must-gather-q82b9"] Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.096457 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ftfvg/must-gather-q82b9" Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.100602 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ftfvg"/"openshift-service-ca.crt" Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.105474 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ftfvg"/"kube-root-ca.crt" Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.127843 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ftfvg/must-gather-q82b9"] Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.216445 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/14200426-4735-4187-b930-33baed713d72-must-gather-output\") pod \"must-gather-q82b9\" (UID: \"14200426-4735-4187-b930-33baed713d72\") " pod="openshift-must-gather-ftfvg/must-gather-q82b9" Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.216543 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7dwp\" (UniqueName: \"kubernetes.io/projected/14200426-4735-4187-b930-33baed713d72-kube-api-access-q7dwp\") pod \"must-gather-q82b9\" (UID: \"14200426-4735-4187-b930-33baed713d72\") " pod="openshift-must-gather-ftfvg/must-gather-q82b9" Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.318519 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/14200426-4735-4187-b930-33baed713d72-must-gather-output\") pod \"must-gather-q82b9\" (UID: \"14200426-4735-4187-b930-33baed713d72\") " pod="openshift-must-gather-ftfvg/must-gather-q82b9" Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.318830 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7dwp\" (UniqueName: \"kubernetes.io/projected/14200426-4735-4187-b930-33baed713d72-kube-api-access-q7dwp\") pod \"must-gather-q82b9\" (UID: \"14200426-4735-4187-b930-33baed713d72\") " pod="openshift-must-gather-ftfvg/must-gather-q82b9" Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.320098 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/14200426-4735-4187-b930-33baed713d72-must-gather-output\") pod \"must-gather-q82b9\" (UID: \"14200426-4735-4187-b930-33baed713d72\") " pod="openshift-must-gather-ftfvg/must-gather-q82b9" Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.344928 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7dwp\" (UniqueName: \"kubernetes.io/projected/14200426-4735-4187-b930-33baed713d72-kube-api-access-q7dwp\") pod \"must-gather-q82b9\" (UID: \"14200426-4735-4187-b930-33baed713d72\") " pod="openshift-must-gather-ftfvg/must-gather-q82b9" Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.420253 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ftfvg/must-gather-q82b9" Feb 18 15:39:25 crc kubenswrapper[4968]: I0218 15:39:25.922159 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ftfvg/must-gather-q82b9"] Feb 18 15:39:26 crc kubenswrapper[4968]: I0218 15:39:26.854198 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ftfvg/must-gather-q82b9" event={"ID":"14200426-4735-4187-b930-33baed713d72","Type":"ContainerStarted","Data":"9051d381db3676ee81eb38eefb7dbddcd41d5d3fd5950dedfb58952cd9ea01ac"} Feb 18 15:39:32 crc kubenswrapper[4968]: I0218 15:39:32.897951 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ftfvg/must-gather-q82b9" event={"ID":"14200426-4735-4187-b930-33baed713d72","Type":"ContainerStarted","Data":"026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c"} Feb 18 15:39:32 crc kubenswrapper[4968]: I0218 15:39:32.898446 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ftfvg/must-gather-q82b9" event={"ID":"14200426-4735-4187-b930-33baed713d72","Type":"ContainerStarted","Data":"b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33"} Feb 18 15:39:32 crc kubenswrapper[4968]: I0218 15:39:32.911402 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ftfvg/must-gather-q82b9" podStartSLOduration=1.651191512 podStartE2EDuration="7.911385277s" podCreationTimestamp="2026-02-18 15:39:25 +0000 UTC" firstStartedPulling="2026-02-18 15:39:25.927875976 +0000 UTC m=+1065.313320838" lastFinishedPulling="2026-02-18 15:39:32.188069741 +0000 UTC m=+1071.573514603" observedRunningTime="2026-02-18 15:39:32.910016288 +0000 UTC m=+1072.295461150" watchObservedRunningTime="2026-02-18 15:39:32.911385277 +0000 UTC m=+1072.296830139" Feb 18 15:40:14 crc kubenswrapper[4968]: I0218 15:40:14.367568 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:40:14 crc kubenswrapper[4968]: I0218 15:40:14.368312 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:40:34 crc kubenswrapper[4968]: I0218 15:40:34.679596 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2_a777cdac-6acd-45d2-9262-8c77577fe2b6/util/0.log" Feb 18 15:40:34 crc kubenswrapper[4968]: I0218 15:40:34.846566 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2_a777cdac-6acd-45d2-9262-8c77577fe2b6/util/0.log" Feb 18 15:40:34 crc kubenswrapper[4968]: I0218 15:40:34.901175 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2_a777cdac-6acd-45d2-9262-8c77577fe2b6/pull/0.log" Feb 18 15:40:34 crc kubenswrapper[4968]: I0218 15:40:34.912364 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2_a777cdac-6acd-45d2-9262-8c77577fe2b6/pull/0.log" Feb 18 15:40:35 crc kubenswrapper[4968]: I0218 15:40:35.033595 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2_a777cdac-6acd-45d2-9262-8c77577fe2b6/util/0.log" Feb 18 15:40:35 crc kubenswrapper[4968]: I0218 15:40:35.058221 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2_a777cdac-6acd-45d2-9262-8c77577fe2b6/pull/0.log" Feb 18 15:40:35 crc kubenswrapper[4968]: I0218 15:40:35.084702 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_74cfd4d5b1f2da5b20c9d31d85f5dbfec7859e4dc0ff9b28c2f46cc04dksvd2_a777cdac-6acd-45d2-9262-8c77577fe2b6/extract/0.log" Feb 18 15:40:35 crc kubenswrapper[4968]: I0218 15:40:35.424470 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-zc86v_87fee87a-b2aa-46b7-8f11-89b00fe7638c/manager/0.log" Feb 18 15:40:35 crc kubenswrapper[4968]: I0218 15:40:35.617930 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-7ltz5_99b546dc-24de-4e98-9fce-7d457bf981ec/manager/0.log" Feb 18 15:40:35 crc kubenswrapper[4968]: I0218 15:40:35.831730 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-kkzx6_cbc0c461-56e5-4386-bb50-a633b89911e0/manager/0.log" Feb 18 15:40:35 crc kubenswrapper[4968]: I0218 15:40:35.904420 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-8ng62_35c5d058-dc84-4faf-a645-c8463d0e6eb1/manager/0.log" Feb 18 15:40:36 crc kubenswrapper[4968]: I0218 15:40:36.059657 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-v9gsm_cae81ff3-9955-429b-9e24-3093d3c7f1fa/manager/0.log" Feb 18 15:40:36 crc kubenswrapper[4968]: I0218 15:40:36.137997 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-lwn4x_fb1b11dc-0a5b-4784-bc3b-a0976c481831/manager/0.log" Feb 18 15:40:36 crc kubenswrapper[4968]: I0218 15:40:36.329629 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-56xzj_6f7c68fd-afbb-4751-b1da-d27bfe32d6c7/manager/0.log" Feb 18 15:40:36 crc kubenswrapper[4968]: I0218 15:40:36.481121 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-6jkdl_60510787-a5c0-467a-969f-38c6be9b2184/manager/0.log" Feb 18 15:40:36 crc kubenswrapper[4968]: I0218 15:40:36.570229 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-f4prf_debbb308-3378-47b6-8e2a-b0d8edc95bfc/manager/0.log" Feb 18 15:40:36 crc kubenswrapper[4968]: I0218 15:40:36.767903 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-mbn84_00926e2b-9085-412d-9b65-2c71f76bd9ed/manager/0.log" Feb 18 15:40:36 crc kubenswrapper[4968]: I0218 15:40:36.902519 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-vvqqh_1bee3920-4de9-436c-8d80-4ea7f0903f11/manager/0.log" Feb 18 15:40:36 crc kubenswrapper[4968]: I0218 15:40:36.950044 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-xhwzj_aa54cd8f-435a-4055-9668-1ab7c3e72516/manager/0.log" Feb 18 15:40:37 crc kubenswrapper[4968]: I0218 15:40:37.529922 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9crpdbz_c2c93869-c328-4b28-8098-18191433d4d2/manager/0.log" Feb 18 15:40:37 crc kubenswrapper[4968]: I0218 15:40:37.883325 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-77859ccd85-qfms4_40e6b6c3-28b7-4626-aff6-5b4932a65ec7/operator/0.log" Feb 18 15:40:37 crc kubenswrapper[4968]: I0218 15:40:37.971324 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-dfg2t_8a5aaad5-3e01-4de2-9bf1-e64680ce7b18/manager/0.log" Feb 18 15:40:38 crc kubenswrapper[4968]: I0218 15:40:38.064185 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-79f9cfbcf7-bzx4g_5e48a888-178c-4d24-b960-60b6251940db/manager/0.log" Feb 18 15:40:38 crc kubenswrapper[4968]: I0218 15:40:38.160292 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-b6km5_858314b2-7c36-467c-a1a5-dd764bd22784/registry-server/0.log" Feb 18 15:40:38 crc kubenswrapper[4968]: I0218 15:40:38.346786 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-jhq7j_53297588-6688-444d-952c-810b9b265acf/manager/0.log" Feb 18 15:40:38 crc kubenswrapper[4968]: I0218 15:40:38.481631 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-wbp4k_964d8817-99e3-4a8d-91ab-876ec187f62d/manager/0.log" Feb 18 15:40:38 crc kubenswrapper[4968]: I0218 15:40:38.597094 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-98jxg_db247d6e-0199-4fde-85d9-86b08817a070/operator/0.log" Feb 18 15:40:38 crc kubenswrapper[4968]: I0218 15:40:38.615957 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-7rw4c_392e8c65-7168-4e8b-b672-fb0b53da6ad5/manager/0.log" Feb 18 15:40:38 crc kubenswrapper[4968]: I0218 15:40:38.704562 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-n6bn4_e39cd00d-1815-442a-9343-277280cf8ccc/manager/0.log" Feb 18 15:40:38 crc kubenswrapper[4968]: I0218 15:40:38.793819 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-kskjh_095020a1-2516-42f1-9223-339fd943c52f/manager/0.log" Feb 18 15:40:38 crc kubenswrapper[4968]: I0218 15:40:38.851618 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-qphmd_dc5fcae4-29ef-4706-9b51-e4f3e99f1c36/manager/0.log" Feb 18 15:40:38 crc kubenswrapper[4968]: I0218 15:40:38.949186 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5c48998cbf-jdwxz_56f5f54a-9f54-4e6e-a9fc-129398f0df0b/manager/0.log" Feb 18 15:40:44 crc kubenswrapper[4968]: I0218 15:40:44.366953 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:40:44 crc kubenswrapper[4968]: I0218 15:40:44.367288 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:40:57 crc kubenswrapper[4968]: I0218 15:40:57.706226 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-x9mtt_605d503b-946e-47e6-b08f-c6eaf0c3eb9a/control-plane-machine-set-operator/0.log" Feb 18 15:40:57 crc kubenswrapper[4968]: I0218 15:40:57.884737 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-h58tr_402543c2-2831-4d3b-87ca-6d284c6d9be1/kube-rbac-proxy/0.log" Feb 18 15:40:57 crc kubenswrapper[4968]: I0218 15:40:57.912214 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-h58tr_402543c2-2831-4d3b-87ca-6d284c6d9be1/machine-api-operator/0.log" Feb 18 15:41:11 crc kubenswrapper[4968]: I0218 15:41:11.113504 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-98hkr_b2fbab5c-3c83-4630-bfba-48bd448b5449/cert-manager-controller/0.log" Feb 18 15:41:11 crc kubenswrapper[4968]: I0218 15:41:11.319144 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-x7mcq_f34e4121-0b18-48e0-b48e-54054ebfbc72/cert-manager-cainjector/0.log" Feb 18 15:41:11 crc kubenswrapper[4968]: I0218 15:41:11.360258 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-wpkq2_30349f7d-e8d4-4791-8668-29e3213e3be3/cert-manager-webhook/0.log" Feb 18 15:41:14 crc kubenswrapper[4968]: I0218 15:41:14.367174 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:41:14 crc kubenswrapper[4968]: I0218 15:41:14.367623 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:41:14 crc kubenswrapper[4968]: I0218 15:41:14.367674 4968 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:41:14 crc kubenswrapper[4968]: I0218 15:41:14.368250 4968 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c0ee8bcfd29be60e6cb05f3e8eb4433c41144c8013553fdaf58be5bb39a80b9"} pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:41:14 crc kubenswrapper[4968]: I0218 15:41:14.368307 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" containerID="cri-o://1c0ee8bcfd29be60e6cb05f3e8eb4433c41144c8013553fdaf58be5bb39a80b9" gracePeriod=600 Feb 18 15:41:14 crc kubenswrapper[4968]: I0218 15:41:14.584826 4968 generic.go:334] "Generic (PLEG): container finished" podID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerID="1c0ee8bcfd29be60e6cb05f3e8eb4433c41144c8013553fdaf58be5bb39a80b9" exitCode=0 Feb 18 15:41:14 crc kubenswrapper[4968]: I0218 15:41:14.584875 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerDied","Data":"1c0ee8bcfd29be60e6cb05f3e8eb4433c41144c8013553fdaf58be5bb39a80b9"} Feb 18 15:41:14 crc kubenswrapper[4968]: I0218 15:41:14.584927 4968 scope.go:117] "RemoveContainer" containerID="2024fbd9a3a558a0c7ebda43f30fc1d9386f6d00027774f8f91dd895439fc899" Feb 18 15:41:15 crc kubenswrapper[4968]: I0218 15:41:15.593713 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerStarted","Data":"79a4ac6e0d12350981bce67b06b3c00e9895b13a7171aa9010c1bcf9c6a8723e"} Feb 18 15:41:23 crc kubenswrapper[4968]: I0218 15:41:23.450395 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-prmzd_3af661c9-d7d5-4b18-8b4a-d6dac9565b6c/nmstate-console-plugin/0.log" Feb 18 15:41:23 crc kubenswrapper[4968]: I0218 15:41:23.602915 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-wsb68_fbb181f8-e6e8-48c4-9598-30622c5e04d8/nmstate-handler/0.log" Feb 18 15:41:23 crc kubenswrapper[4968]: I0218 15:41:23.628385 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-cfdkm_e460ba68-eeb3-4979-a32d-8c2866dfbdb9/nmstate-metrics/0.log" Feb 18 15:41:23 crc kubenswrapper[4968]: I0218 15:41:23.659421 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-cfdkm_e460ba68-eeb3-4979-a32d-8c2866dfbdb9/kube-rbac-proxy/0.log" Feb 18 15:41:23 crc kubenswrapper[4968]: I0218 15:41:23.785947 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-jlkfr_63e09cfe-38d6-44fd-859c-3bc876d8808f/nmstate-operator/0.log" Feb 18 15:41:23 crc kubenswrapper[4968]: I0218 15:41:23.842640 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-zwksb_81a82099-84e4-40c8-8dd0-b8803001bebf/nmstate-webhook/0.log" Feb 18 15:41:37 crc kubenswrapper[4968]: I0218 15:41:37.525349 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-d8lx8_96458238-fdec-497e-988d-0d6d6a59af36/prometheus-operator/0.log" Feb 18 15:41:37 crc kubenswrapper[4968]: I0218 15:41:37.792630 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw_de5c536d-8040-4425-8bf1-5deda35fe0c6/prometheus-operator-admission-webhook/0.log" Feb 18 15:41:37 crc kubenswrapper[4968]: I0218 15:41:37.859989 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-57669fbcb7-8777x_d21d7e50-1c72-4fc6-ad51-3603cccbce41/prometheus-operator-admission-webhook/0.log" Feb 18 15:41:37 crc kubenswrapper[4968]: I0218 15:41:37.995789 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-mg5x7_c1f3704e-f185-49f7-8d17-226dd2f42061/operator/0.log" Feb 18 15:41:38 crc kubenswrapper[4968]: I0218 15:41:38.091889 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-7f95q_ff69eec4-70ef-4b36-ab53-93131c8d15ce/perses-operator/0.log" Feb 18 15:41:52 crc kubenswrapper[4968]: I0218 15:41:52.242822 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-7ghrx_7ab9fc6e-f33c-4351-b19a-638d53946dd0/kube-rbac-proxy/0.log" Feb 18 15:41:52 crc kubenswrapper[4968]: I0218 15:41:52.257482 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-7ghrx_7ab9fc6e-f33c-4351-b19a-638d53946dd0/controller/0.log" Feb 18 15:41:52 crc kubenswrapper[4968]: I0218 15:41:52.414297 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/cp-frr-files/0.log" Feb 18 15:41:52 crc kubenswrapper[4968]: I0218 15:41:52.605592 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/cp-frr-files/0.log" Feb 18 15:41:52 crc kubenswrapper[4968]: I0218 15:41:52.612050 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/cp-reloader/0.log" Feb 18 15:41:52 crc kubenswrapper[4968]: I0218 15:41:52.631391 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/cp-reloader/0.log" Feb 18 15:41:52 crc kubenswrapper[4968]: I0218 15:41:52.644410 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/cp-metrics/0.log" Feb 18 15:41:52 crc kubenswrapper[4968]: I0218 15:41:52.871772 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/cp-frr-files/0.log" Feb 18 15:41:52 crc kubenswrapper[4968]: I0218 15:41:52.880965 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/cp-metrics/0.log" Feb 18 15:41:52 crc kubenswrapper[4968]: I0218 15:41:52.888758 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/cp-reloader/0.log" Feb 18 15:41:52 crc kubenswrapper[4968]: I0218 15:41:52.943684 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/cp-metrics/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.100676 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/cp-frr-files/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.108515 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/cp-metrics/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.112168 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/cp-reloader/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.125838 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/controller/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.264984 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/frr-metrics/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.295189 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/kube-rbac-proxy/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.333402 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/kube-rbac-proxy-frr/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.396517 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/frr/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.512939 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-nz96q_56af3b3d-b4ba-4fe9-832f-969d3a767ed2/reloader/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.551056 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-7c58z_19821749-39d5-4ca0-8fb0-3b7edca2da5a/frr-k8s-webhook-server/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.646386 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-f8cdcf667-bh7xq_4e91721f-0e20-4dfb-9e17-45107d5cf8ce/manager/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.749007 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5c865dc449-g7xl7_8787b83e-ce8e-4819-84f2-08adb1f238c8/webhook-server/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.830458 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-z4qpq_d3264f33-404d-4edb-a17a-1273abc843b5/kube-rbac-proxy/0.log" Feb 18 15:41:53 crc kubenswrapper[4968]: I0218 15:41:53.961908 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-z4qpq_d3264f33-404d-4edb-a17a-1273abc843b5/speaker/0.log" Feb 18 15:42:08 crc kubenswrapper[4968]: I0218 15:42:08.286667 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_62b19fba-8472-4001-b221-2d4f0dd1d36e/util/0.log" Feb 18 15:42:08 crc kubenswrapper[4968]: I0218 15:42:08.511041 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_62b19fba-8472-4001-b221-2d4f0dd1d36e/util/0.log" Feb 18 15:42:08 crc kubenswrapper[4968]: I0218 15:42:08.544251 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_62b19fba-8472-4001-b221-2d4f0dd1d36e/pull/0.log" Feb 18 15:42:08 crc kubenswrapper[4968]: I0218 15:42:08.544392 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_62b19fba-8472-4001-b221-2d4f0dd1d36e/pull/0.log" Feb 18 15:42:08 crc kubenswrapper[4968]: I0218 15:42:08.818555 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_62b19fba-8472-4001-b221-2d4f0dd1d36e/pull/0.log" Feb 18 15:42:08 crc kubenswrapper[4968]: I0218 15:42:08.829062 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_62b19fba-8472-4001-b221-2d4f0dd1d36e/extract/0.log" Feb 18 15:42:08 crc kubenswrapper[4968]: I0218 15:42:08.858614 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f085thv8_62b19fba-8472-4001-b221-2d4f0dd1d36e/util/0.log" Feb 18 15:42:08 crc kubenswrapper[4968]: I0218 15:42:08.989024 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh_e034e9a0-5038-41d3-8372-78e611d8b952/util/0.log" Feb 18 15:42:09 crc kubenswrapper[4968]: I0218 15:42:09.159805 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh_e034e9a0-5038-41d3-8372-78e611d8b952/pull/0.log" Feb 18 15:42:09 crc kubenswrapper[4968]: I0218 15:42:09.162106 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh_e034e9a0-5038-41d3-8372-78e611d8b952/util/0.log" Feb 18 15:42:09 crc kubenswrapper[4968]: I0218 15:42:09.164861 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh_e034e9a0-5038-41d3-8372-78e611d8b952/pull/0.log" Feb 18 15:42:09 crc kubenswrapper[4968]: I0218 15:42:09.314601 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh_e034e9a0-5038-41d3-8372-78e611d8b952/pull/0.log" Feb 18 15:42:09 crc kubenswrapper[4968]: I0218 15:42:09.328236 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh_e034e9a0-5038-41d3-8372-78e611d8b952/util/0.log" Feb 18 15:42:09 crc kubenswrapper[4968]: I0218 15:42:09.357344 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2134w8wh_e034e9a0-5038-41d3-8372-78e611d8b952/extract/0.log" Feb 18 15:42:09 crc kubenswrapper[4968]: I0218 15:42:09.497854 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-blvt2_3589f4c2-2e33-4458-ae09-8844e66b94fd/extract-utilities/0.log" Feb 18 15:42:09 crc kubenswrapper[4968]: I0218 15:42:09.677274 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-blvt2_3589f4c2-2e33-4458-ae09-8844e66b94fd/extract-content/0.log" Feb 18 15:42:09 crc kubenswrapper[4968]: I0218 15:42:09.683420 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-blvt2_3589f4c2-2e33-4458-ae09-8844e66b94fd/extract-content/0.log" Feb 18 15:42:09 crc kubenswrapper[4968]: I0218 15:42:09.683842 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-blvt2_3589f4c2-2e33-4458-ae09-8844e66b94fd/extract-utilities/0.log" Feb 18 15:42:09 crc kubenswrapper[4968]: I0218 15:42:09.814928 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-blvt2_3589f4c2-2e33-4458-ae09-8844e66b94fd/extract-utilities/0.log" Feb 18 15:42:09 crc kubenswrapper[4968]: I0218 15:42:09.860614 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-blvt2_3589f4c2-2e33-4458-ae09-8844e66b94fd/extract-content/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.015884 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4kpft_5391b9a3-b12d-4c69-9d11-b29cbc44b531/extract-utilities/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.097353 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-blvt2_3589f4c2-2e33-4458-ae09-8844e66b94fd/registry-server/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.159120 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4kpft_5391b9a3-b12d-4c69-9d11-b29cbc44b531/extract-content/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.201149 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4kpft_5391b9a3-b12d-4c69-9d11-b29cbc44b531/extract-utilities/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.218202 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4kpft_5391b9a3-b12d-4c69-9d11-b29cbc44b531/extract-content/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.322202 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4kpft_5391b9a3-b12d-4c69-9d11-b29cbc44b531/extract-utilities/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.330014 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4kpft_5391b9a3-b12d-4c69-9d11-b29cbc44b531/extract-content/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.564577 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k_67350a11-27a5-46e6-bab7-32c8a7bd74df/util/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.685388 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k_67350a11-27a5-46e6-bab7-32c8a7bd74df/util/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.699565 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4kpft_5391b9a3-b12d-4c69-9d11-b29cbc44b531/registry-server/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.728482 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k_67350a11-27a5-46e6-bab7-32c8a7bd74df/pull/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.774973 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k_67350a11-27a5-46e6-bab7-32c8a7bd74df/pull/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.948690 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k_67350a11-27a5-46e6-bab7-32c8a7bd74df/util/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.973420 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k_67350a11-27a5-46e6-bab7-32c8a7bd74df/extract/0.log" Feb 18 15:42:10 crc kubenswrapper[4968]: I0218 15:42:10.980151 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecagk24k_67350a11-27a5-46e6-bab7-32c8a7bd74df/pull/0.log" Feb 18 15:42:11 crc kubenswrapper[4968]: I0218 15:42:11.123142 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8m6ss_c09f5ccb-47bd-46ca-8f51-1fc196f396ee/extract-utilities/0.log" Feb 18 15:42:11 crc kubenswrapper[4968]: I0218 15:42:11.144706 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-242v6_e2eb1519-86a1-496a-ace0-ee997d43287c/marketplace-operator/0.log" Feb 18 15:42:11 crc kubenswrapper[4968]: I0218 15:42:11.303628 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8m6ss_c09f5ccb-47bd-46ca-8f51-1fc196f396ee/extract-utilities/0.log" Feb 18 15:42:11 crc kubenswrapper[4968]: I0218 15:42:11.306606 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8m6ss_c09f5ccb-47bd-46ca-8f51-1fc196f396ee/extract-content/0.log" Feb 18 15:42:11 crc kubenswrapper[4968]: I0218 15:42:11.350268 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8m6ss_c09f5ccb-47bd-46ca-8f51-1fc196f396ee/extract-content/0.log" Feb 18 15:42:11 crc kubenswrapper[4968]: I0218 15:42:11.499606 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8m6ss_c09f5ccb-47bd-46ca-8f51-1fc196f396ee/extract-content/0.log" Feb 18 15:42:11 crc kubenswrapper[4968]: I0218 15:42:11.531172 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8m6ss_c09f5ccb-47bd-46ca-8f51-1fc196f396ee/extract-utilities/0.log" Feb 18 15:42:11 crc kubenswrapper[4968]: I0218 15:42:11.578587 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8m6ss_c09f5ccb-47bd-46ca-8f51-1fc196f396ee/registry-server/0.log" Feb 18 15:42:11 crc kubenswrapper[4968]: I0218 15:42:11.687781 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9v9wz_d3c68c1d-979b-4059-96bb-8c1415fb9776/extract-utilities/0.log" Feb 18 15:42:11 crc kubenswrapper[4968]: I0218 15:42:11.865788 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9v9wz_d3c68c1d-979b-4059-96bb-8c1415fb9776/extract-content/0.log" Feb 18 15:42:11 crc kubenswrapper[4968]: I0218 15:42:11.875625 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9v9wz_d3c68c1d-979b-4059-96bb-8c1415fb9776/extract-utilities/0.log" Feb 18 15:42:11 crc kubenswrapper[4968]: I0218 15:42:11.916914 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9v9wz_d3c68c1d-979b-4059-96bb-8c1415fb9776/extract-content/0.log" Feb 18 15:42:12 crc kubenswrapper[4968]: I0218 15:42:12.040651 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9v9wz_d3c68c1d-979b-4059-96bb-8c1415fb9776/extract-utilities/0.log" Feb 18 15:42:12 crc kubenswrapper[4968]: I0218 15:42:12.068620 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9v9wz_d3c68c1d-979b-4059-96bb-8c1415fb9776/extract-content/0.log" Feb 18 15:42:12 crc kubenswrapper[4968]: I0218 15:42:12.342023 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9v9wz_d3c68c1d-979b-4059-96bb-8c1415fb9776/registry-server/0.log" Feb 18 15:42:24 crc kubenswrapper[4968]: I0218 15:42:24.735520 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-57669fbcb7-45bdw_de5c536d-8040-4425-8bf1-5deda35fe0c6/prometheus-operator-admission-webhook/0.log" Feb 18 15:42:24 crc kubenswrapper[4968]: I0218 15:42:24.796918 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-d8lx8_96458238-fdec-497e-988d-0d6d6a59af36/prometheus-operator/0.log" Feb 18 15:42:24 crc kubenswrapper[4968]: I0218 15:42:24.816151 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-57669fbcb7-8777x_d21d7e50-1c72-4fc6-ad51-3603cccbce41/prometheus-operator-admission-webhook/0.log" Feb 18 15:42:24 crc kubenswrapper[4968]: I0218 15:42:24.929955 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-mg5x7_c1f3704e-f185-49f7-8d17-226dd2f42061/operator/0.log" Feb 18 15:42:24 crc kubenswrapper[4968]: I0218 15:42:24.992147 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-7f95q_ff69eec4-70ef-4b36-ab53-93131c8d15ce/perses-operator/0.log" Feb 18 15:43:14 crc kubenswrapper[4968]: I0218 15:43:14.367529 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:43:14 crc kubenswrapper[4968]: I0218 15:43:14.368100 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:43:22 crc kubenswrapper[4968]: I0218 15:43:22.561286 4968 generic.go:334] "Generic (PLEG): container finished" podID="14200426-4735-4187-b930-33baed713d72" containerID="b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33" exitCode=0 Feb 18 15:43:22 crc kubenswrapper[4968]: I0218 15:43:22.561466 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ftfvg/must-gather-q82b9" event={"ID":"14200426-4735-4187-b930-33baed713d72","Type":"ContainerDied","Data":"b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33"} Feb 18 15:43:22 crc kubenswrapper[4968]: I0218 15:43:22.563215 4968 scope.go:117] "RemoveContainer" containerID="b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33" Feb 18 15:43:22 crc kubenswrapper[4968]: I0218 15:43:22.714234 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ftfvg_must-gather-q82b9_14200426-4735-4187-b930-33baed713d72/gather/0.log" Feb 18 15:43:30 crc kubenswrapper[4968]: I0218 15:43:30.554861 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ftfvg/must-gather-q82b9"] Feb 18 15:43:30 crc kubenswrapper[4968]: I0218 15:43:30.555682 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ftfvg/must-gather-q82b9" podUID="14200426-4735-4187-b930-33baed713d72" containerName="copy" containerID="cri-o://026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c" gracePeriod=2 Feb 18 15:43:30 crc kubenswrapper[4968]: I0218 15:43:30.565503 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ftfvg/must-gather-q82b9"] Feb 18 15:43:30 crc kubenswrapper[4968]: E0218 15:43:30.634009 4968 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14200426_4735_4187_b930_33baed713d72.slice/crio-026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c.scope\": RecentStats: unable to find data in memory cache]" Feb 18 15:43:30 crc kubenswrapper[4968]: I0218 15:43:30.885717 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ftfvg_must-gather-q82b9_14200426-4735-4187-b930-33baed713d72/copy/0.log" Feb 18 15:43:30 crc kubenswrapper[4968]: I0218 15:43:30.886424 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ftfvg/must-gather-q82b9" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.009768 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/14200426-4735-4187-b930-33baed713d72-must-gather-output\") pod \"14200426-4735-4187-b930-33baed713d72\" (UID: \"14200426-4735-4187-b930-33baed713d72\") " Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.009820 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7dwp\" (UniqueName: \"kubernetes.io/projected/14200426-4735-4187-b930-33baed713d72-kube-api-access-q7dwp\") pod \"14200426-4735-4187-b930-33baed713d72\" (UID: \"14200426-4735-4187-b930-33baed713d72\") " Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.017955 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14200426-4735-4187-b930-33baed713d72-kube-api-access-q7dwp" (OuterVolumeSpecName: "kube-api-access-q7dwp") pod "14200426-4735-4187-b930-33baed713d72" (UID: "14200426-4735-4187-b930-33baed713d72"). InnerVolumeSpecName "kube-api-access-q7dwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.110676 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14200426-4735-4187-b930-33baed713d72-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "14200426-4735-4187-b930-33baed713d72" (UID: "14200426-4735-4187-b930-33baed713d72"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.111183 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/14200426-4735-4187-b930-33baed713d72-must-gather-output\") pod \"14200426-4735-4187-b930-33baed713d72\" (UID: \"14200426-4735-4187-b930-33baed713d72\") " Feb 18 15:43:31 crc kubenswrapper[4968]: W0218 15:43:31.111301 4968 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/14200426-4735-4187-b930-33baed713d72/volumes/kubernetes.io~empty-dir/must-gather-output Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.111317 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14200426-4735-4187-b930-33baed713d72-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "14200426-4735-4187-b930-33baed713d72" (UID: "14200426-4735-4187-b930-33baed713d72"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.111555 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7dwp\" (UniqueName: \"kubernetes.io/projected/14200426-4735-4187-b930-33baed713d72-kube-api-access-q7dwp\") on node \"crc\" DevicePath \"\"" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.111588 4968 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/14200426-4735-4187-b930-33baed713d72-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.271912 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14200426-4735-4187-b930-33baed713d72" path="/var/lib/kubelet/pods/14200426-4735-4187-b930-33baed713d72/volumes" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.648228 4968 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ftfvg_must-gather-q82b9_14200426-4735-4187-b930-33baed713d72/copy/0.log" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.648630 4968 generic.go:334] "Generic (PLEG): container finished" podID="14200426-4735-4187-b930-33baed713d72" containerID="026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c" exitCode=143 Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.648692 4968 scope.go:117] "RemoveContainer" containerID="026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.648903 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ftfvg/must-gather-q82b9" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.678223 4968 scope.go:117] "RemoveContainer" containerID="b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.727376 4968 scope.go:117] "RemoveContainer" containerID="026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c" Feb 18 15:43:31 crc kubenswrapper[4968]: E0218 15:43:31.728114 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c\": container with ID starting with 026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c not found: ID does not exist" containerID="026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.728175 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c"} err="failed to get container status \"026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c\": rpc error: code = NotFound desc = could not find container \"026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c\": container with ID starting with 026256a2f6ea592b222f7470b3923e50b5056325eeb9c0cf7d0766a5f2b8ec3c not found: ID does not exist" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.728209 4968 scope.go:117] "RemoveContainer" containerID="b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33" Feb 18 15:43:31 crc kubenswrapper[4968]: E0218 15:43:31.728515 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33\": container with ID starting with b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33 not found: ID does not exist" containerID="b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33" Feb 18 15:43:31 crc kubenswrapper[4968]: I0218 15:43:31.728548 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33"} err="failed to get container status \"b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33\": rpc error: code = NotFound desc = could not find container \"b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33\": container with ID starting with b0f592b47935d3d907f6b38489fd0ee3bb152927d437de4dba9a1681931c5d33 not found: ID does not exist" Feb 18 15:43:44 crc kubenswrapper[4968]: I0218 15:43:44.367167 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:43:44 crc kubenswrapper[4968]: I0218 15:43:44.367966 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:44:14 crc kubenswrapper[4968]: I0218 15:44:14.367710 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:44:14 crc kubenswrapper[4968]: I0218 15:44:14.368469 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:44:14 crc kubenswrapper[4968]: I0218 15:44:14.368532 4968 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:44:14 crc kubenswrapper[4968]: I0218 15:44:14.369328 4968 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"79a4ac6e0d12350981bce67b06b3c00e9895b13a7171aa9010c1bcf9c6a8723e"} pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:44:14 crc kubenswrapper[4968]: I0218 15:44:14.369410 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" containerID="cri-o://79a4ac6e0d12350981bce67b06b3c00e9895b13a7171aa9010c1bcf9c6a8723e" gracePeriod=600 Feb 18 15:44:15 crc kubenswrapper[4968]: I0218 15:44:15.016900 4968 generic.go:334] "Generic (PLEG): container finished" podID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerID="79a4ac6e0d12350981bce67b06b3c00e9895b13a7171aa9010c1bcf9c6a8723e" exitCode=0 Feb 18 15:44:15 crc kubenswrapper[4968]: I0218 15:44:15.017810 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerDied","Data":"79a4ac6e0d12350981bce67b06b3c00e9895b13a7171aa9010c1bcf9c6a8723e"} Feb 18 15:44:15 crc kubenswrapper[4968]: I0218 15:44:15.017853 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerStarted","Data":"a5f596668788929acd43efc0faf07605f488d17a5ceabc4b158d494ee696c99c"} Feb 18 15:44:15 crc kubenswrapper[4968]: I0218 15:44:15.017880 4968 scope.go:117] "RemoveContainer" containerID="1c0ee8bcfd29be60e6cb05f3e8eb4433c41144c8013553fdaf58be5bb39a80b9" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.187670 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cxh2q"] Feb 18 15:44:33 crc kubenswrapper[4968]: E0218 15:44:33.188717 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14200426-4735-4187-b930-33baed713d72" containerName="gather" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.188743 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="14200426-4735-4187-b930-33baed713d72" containerName="gather" Feb 18 15:44:33 crc kubenswrapper[4968]: E0218 15:44:33.188828 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14200426-4735-4187-b930-33baed713d72" containerName="copy" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.188840 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="14200426-4735-4187-b930-33baed713d72" containerName="copy" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.189071 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="14200426-4735-4187-b930-33baed713d72" containerName="gather" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.189104 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="14200426-4735-4187-b930-33baed713d72" containerName="copy" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.190659 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.202519 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxh2q"] Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.268352 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwrhr\" (UniqueName: \"kubernetes.io/projected/a59d494a-c4b6-4673-a9a1-f85100b52ef0-kube-api-access-fwrhr\") pod \"redhat-marketplace-cxh2q\" (UID: \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\") " pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.268508 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a59d494a-c4b6-4673-a9a1-f85100b52ef0-utilities\") pod \"redhat-marketplace-cxh2q\" (UID: \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\") " pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.268545 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a59d494a-c4b6-4673-a9a1-f85100b52ef0-catalog-content\") pod \"redhat-marketplace-cxh2q\" (UID: \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\") " pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.369682 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a59d494a-c4b6-4673-a9a1-f85100b52ef0-catalog-content\") pod \"redhat-marketplace-cxh2q\" (UID: \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\") " pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.369732 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a59d494a-c4b6-4673-a9a1-f85100b52ef0-utilities\") pod \"redhat-marketplace-cxh2q\" (UID: \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\") " pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.369828 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwrhr\" (UniqueName: \"kubernetes.io/projected/a59d494a-c4b6-4673-a9a1-f85100b52ef0-kube-api-access-fwrhr\") pod \"redhat-marketplace-cxh2q\" (UID: \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\") " pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.370290 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a59d494a-c4b6-4673-a9a1-f85100b52ef0-catalog-content\") pod \"redhat-marketplace-cxh2q\" (UID: \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\") " pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.370293 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a59d494a-c4b6-4673-a9a1-f85100b52ef0-utilities\") pod \"redhat-marketplace-cxh2q\" (UID: \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\") " pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.390188 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwrhr\" (UniqueName: \"kubernetes.io/projected/a59d494a-c4b6-4673-a9a1-f85100b52ef0-kube-api-access-fwrhr\") pod \"redhat-marketplace-cxh2q\" (UID: \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\") " pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.512826 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.787363 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w2sfw"] Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.789215 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.798040 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w2sfw"] Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.875983 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95cc5d12-9b01-4d94-be96-8268bc5e6bd3-catalog-content\") pod \"redhat-operators-w2sfw\" (UID: \"95cc5d12-9b01-4d94-be96-8268bc5e6bd3\") " pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.876222 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wjgg\" (UniqueName: \"kubernetes.io/projected/95cc5d12-9b01-4d94-be96-8268bc5e6bd3-kube-api-access-5wjgg\") pod \"redhat-operators-w2sfw\" (UID: \"95cc5d12-9b01-4d94-be96-8268bc5e6bd3\") " pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.876407 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95cc5d12-9b01-4d94-be96-8268bc5e6bd3-utilities\") pod \"redhat-operators-w2sfw\" (UID: \"95cc5d12-9b01-4d94-be96-8268bc5e6bd3\") " pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.978309 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95cc5d12-9b01-4d94-be96-8268bc5e6bd3-utilities\") pod \"redhat-operators-w2sfw\" (UID: \"95cc5d12-9b01-4d94-be96-8268bc5e6bd3\") " pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.978385 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95cc5d12-9b01-4d94-be96-8268bc5e6bd3-catalog-content\") pod \"redhat-operators-w2sfw\" (UID: \"95cc5d12-9b01-4d94-be96-8268bc5e6bd3\") " pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.978453 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wjgg\" (UniqueName: \"kubernetes.io/projected/95cc5d12-9b01-4d94-be96-8268bc5e6bd3-kube-api-access-5wjgg\") pod \"redhat-operators-w2sfw\" (UID: \"95cc5d12-9b01-4d94-be96-8268bc5e6bd3\") " pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.978899 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95cc5d12-9b01-4d94-be96-8268bc5e6bd3-utilities\") pod \"redhat-operators-w2sfw\" (UID: \"95cc5d12-9b01-4d94-be96-8268bc5e6bd3\") " pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.979028 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95cc5d12-9b01-4d94-be96-8268bc5e6bd3-catalog-content\") pod \"redhat-operators-w2sfw\" (UID: \"95cc5d12-9b01-4d94-be96-8268bc5e6bd3\") " pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:33 crc kubenswrapper[4968]: I0218 15:44:33.986504 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxh2q"] Feb 18 15:44:34 crc kubenswrapper[4968]: I0218 15:44:34.004240 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wjgg\" (UniqueName: \"kubernetes.io/projected/95cc5d12-9b01-4d94-be96-8268bc5e6bd3-kube-api-access-5wjgg\") pod \"redhat-operators-w2sfw\" (UID: \"95cc5d12-9b01-4d94-be96-8268bc5e6bd3\") " pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:34 crc kubenswrapper[4968]: I0218 15:44:34.113561 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:34 crc kubenswrapper[4968]: I0218 15:44:34.190529 4968 generic.go:334] "Generic (PLEG): container finished" podID="a59d494a-c4b6-4673-a9a1-f85100b52ef0" containerID="1e31eb7ea8870715cf7f2e37b7cb29168240b9af6bd97361e9b258ce927a6e4e" exitCode=0 Feb 18 15:44:34 crc kubenswrapper[4968]: I0218 15:44:34.190569 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxh2q" event={"ID":"a59d494a-c4b6-4673-a9a1-f85100b52ef0","Type":"ContainerDied","Data":"1e31eb7ea8870715cf7f2e37b7cb29168240b9af6bd97361e9b258ce927a6e4e"} Feb 18 15:44:34 crc kubenswrapper[4968]: I0218 15:44:34.190594 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxh2q" event={"ID":"a59d494a-c4b6-4673-a9a1-f85100b52ef0","Type":"ContainerStarted","Data":"6ad737b5ff9175c96106e73147f49269a82c40093ffec232fcbd67ba9fe31ea0"} Feb 18 15:44:34 crc kubenswrapper[4968]: I0218 15:44:34.192821 4968 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 15:44:34 crc kubenswrapper[4968]: I0218 15:44:34.346499 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w2sfw"] Feb 18 15:44:35 crc kubenswrapper[4968]: I0218 15:44:35.198826 4968 generic.go:334] "Generic (PLEG): container finished" podID="95cc5d12-9b01-4d94-be96-8268bc5e6bd3" containerID="3d2f6b3ce2bc1e49f08079c3bd068902d92f3c6f806bdbc2c72ae03c9d244a37" exitCode=0 Feb 18 15:44:35 crc kubenswrapper[4968]: I0218 15:44:35.198962 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2sfw" event={"ID":"95cc5d12-9b01-4d94-be96-8268bc5e6bd3","Type":"ContainerDied","Data":"3d2f6b3ce2bc1e49f08079c3bd068902d92f3c6f806bdbc2c72ae03c9d244a37"} Feb 18 15:44:35 crc kubenswrapper[4968]: I0218 15:44:35.199265 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2sfw" event={"ID":"95cc5d12-9b01-4d94-be96-8268bc5e6bd3","Type":"ContainerStarted","Data":"da89907301d1a5d289ce31474b0ca80b2371784fd9bf4198073a5692278f7c83"} Feb 18 15:44:35 crc kubenswrapper[4968]: I0218 15:44:35.204534 4968 generic.go:334] "Generic (PLEG): container finished" podID="a59d494a-c4b6-4673-a9a1-f85100b52ef0" containerID="11cab75071b70b8832e40c7683f03cc3798351958a69d001e69f6d4392e85adc" exitCode=0 Feb 18 15:44:35 crc kubenswrapper[4968]: I0218 15:44:35.204570 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxh2q" event={"ID":"a59d494a-c4b6-4673-a9a1-f85100b52ef0","Type":"ContainerDied","Data":"11cab75071b70b8832e40c7683f03cc3798351958a69d001e69f6d4392e85adc"} Feb 18 15:44:36 crc kubenswrapper[4968]: I0218 15:44:36.211857 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxh2q" event={"ID":"a59d494a-c4b6-4673-a9a1-f85100b52ef0","Type":"ContainerStarted","Data":"f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b"} Feb 18 15:44:36 crc kubenswrapper[4968]: I0218 15:44:36.229263 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cxh2q" podStartSLOduration=1.793649593 podStartE2EDuration="3.229244276s" podCreationTimestamp="2026-02-18 15:44:33 +0000 UTC" firstStartedPulling="2026-02-18 15:44:34.19253581 +0000 UTC m=+1373.577980672" lastFinishedPulling="2026-02-18 15:44:35.628130493 +0000 UTC m=+1375.013575355" observedRunningTime="2026-02-18 15:44:36.227740193 +0000 UTC m=+1375.613185055" watchObservedRunningTime="2026-02-18 15:44:36.229244276 +0000 UTC m=+1375.614689138" Feb 18 15:44:43 crc kubenswrapper[4968]: I0218 15:44:43.266909 4968 generic.go:334] "Generic (PLEG): container finished" podID="95cc5d12-9b01-4d94-be96-8268bc5e6bd3" containerID="d35e641fe128e7fcb94d3db40df4a5eae68e564561691b6994b1344b260b4419" exitCode=0 Feb 18 15:44:43 crc kubenswrapper[4968]: I0218 15:44:43.267002 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2sfw" event={"ID":"95cc5d12-9b01-4d94-be96-8268bc5e6bd3","Type":"ContainerDied","Data":"d35e641fe128e7fcb94d3db40df4a5eae68e564561691b6994b1344b260b4419"} Feb 18 15:44:43 crc kubenswrapper[4968]: I0218 15:44:43.513974 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:43 crc kubenswrapper[4968]: I0218 15:44:43.514506 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:43 crc kubenswrapper[4968]: I0218 15:44:43.550963 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:44 crc kubenswrapper[4968]: I0218 15:44:44.277141 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2sfw" event={"ID":"95cc5d12-9b01-4d94-be96-8268bc5e6bd3","Type":"ContainerStarted","Data":"cfecb9c7bb80f67f7ba22ca2148fda9b5d5059cb54c52e9d1078816322d142a1"} Feb 18 15:44:44 crc kubenswrapper[4968]: I0218 15:44:44.299561 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w2sfw" podStartSLOduration=2.512118288 podStartE2EDuration="11.299542982s" podCreationTimestamp="2026-02-18 15:44:33 +0000 UTC" firstStartedPulling="2026-02-18 15:44:35.200821938 +0000 UTC m=+1374.586266800" lastFinishedPulling="2026-02-18 15:44:43.988246622 +0000 UTC m=+1383.373691494" observedRunningTime="2026-02-18 15:44:44.296575258 +0000 UTC m=+1383.682020150" watchObservedRunningTime="2026-02-18 15:44:44.299542982 +0000 UTC m=+1383.684987854" Feb 18 15:44:44 crc kubenswrapper[4968]: I0218 15:44:44.335196 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:45 crc kubenswrapper[4968]: I0218 15:44:45.497559 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxh2q"] Feb 18 15:44:47 crc kubenswrapper[4968]: I0218 15:44:47.298880 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cxh2q" podUID="a59d494a-c4b6-4673-a9a1-f85100b52ef0" containerName="registry-server" containerID="cri-o://f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b" gracePeriod=2 Feb 18 15:44:47 crc kubenswrapper[4968]: I0218 15:44:47.714243 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:47 crc kubenswrapper[4968]: I0218 15:44:47.782291 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a59d494a-c4b6-4673-a9a1-f85100b52ef0-utilities\") pod \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\" (UID: \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\") " Feb 18 15:44:47 crc kubenswrapper[4968]: I0218 15:44:47.782649 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a59d494a-c4b6-4673-a9a1-f85100b52ef0-catalog-content\") pod \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\" (UID: \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\") " Feb 18 15:44:47 crc kubenswrapper[4968]: I0218 15:44:47.782825 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwrhr\" (UniqueName: \"kubernetes.io/projected/a59d494a-c4b6-4673-a9a1-f85100b52ef0-kube-api-access-fwrhr\") pod \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\" (UID: \"a59d494a-c4b6-4673-a9a1-f85100b52ef0\") " Feb 18 15:44:47 crc kubenswrapper[4968]: I0218 15:44:47.784117 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a59d494a-c4b6-4673-a9a1-f85100b52ef0-utilities" (OuterVolumeSpecName: "utilities") pod "a59d494a-c4b6-4673-a9a1-f85100b52ef0" (UID: "a59d494a-c4b6-4673-a9a1-f85100b52ef0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:44:47 crc kubenswrapper[4968]: I0218 15:44:47.789609 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a59d494a-c4b6-4673-a9a1-f85100b52ef0-kube-api-access-fwrhr" (OuterVolumeSpecName: "kube-api-access-fwrhr") pod "a59d494a-c4b6-4673-a9a1-f85100b52ef0" (UID: "a59d494a-c4b6-4673-a9a1-f85100b52ef0"). InnerVolumeSpecName "kube-api-access-fwrhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:44:47 crc kubenswrapper[4968]: I0218 15:44:47.808622 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a59d494a-c4b6-4673-a9a1-f85100b52ef0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a59d494a-c4b6-4673-a9a1-f85100b52ef0" (UID: "a59d494a-c4b6-4673-a9a1-f85100b52ef0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:44:47 crc kubenswrapper[4968]: I0218 15:44:47.884471 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a59d494a-c4b6-4673-a9a1-f85100b52ef0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:44:47 crc kubenswrapper[4968]: I0218 15:44:47.884519 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwrhr\" (UniqueName: \"kubernetes.io/projected/a59d494a-c4b6-4673-a9a1-f85100b52ef0-kube-api-access-fwrhr\") on node \"crc\" DevicePath \"\"" Feb 18 15:44:47 crc kubenswrapper[4968]: I0218 15:44:47.884541 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a59d494a-c4b6-4673-a9a1-f85100b52ef0-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.309289 4968 generic.go:334] "Generic (PLEG): container finished" podID="a59d494a-c4b6-4673-a9a1-f85100b52ef0" containerID="f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b" exitCode=0 Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.309363 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxh2q" event={"ID":"a59d494a-c4b6-4673-a9a1-f85100b52ef0","Type":"ContainerDied","Data":"f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b"} Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.309388 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxh2q" event={"ID":"a59d494a-c4b6-4673-a9a1-f85100b52ef0","Type":"ContainerDied","Data":"6ad737b5ff9175c96106e73147f49269a82c40093ffec232fcbd67ba9fe31ea0"} Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.309403 4968 scope.go:117] "RemoveContainer" containerID="f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b" Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.311869 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxh2q" Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.331602 4968 scope.go:117] "RemoveContainer" containerID="11cab75071b70b8832e40c7683f03cc3798351958a69d001e69f6d4392e85adc" Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.351177 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxh2q"] Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.358309 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxh2q"] Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.370594 4968 scope.go:117] "RemoveContainer" containerID="1e31eb7ea8870715cf7f2e37b7cb29168240b9af6bd97361e9b258ce927a6e4e" Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.386380 4968 scope.go:117] "RemoveContainer" containerID="f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b" Feb 18 15:44:48 crc kubenswrapper[4968]: E0218 15:44:48.386826 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b\": container with ID starting with f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b not found: ID does not exist" containerID="f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b" Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.386857 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b"} err="failed to get container status \"f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b\": rpc error: code = NotFound desc = could not find container \"f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b\": container with ID starting with f35f646f9b214a4f71ab5e22484e935d1068864119fb19cc2175bfe92d8cb68b not found: ID does not exist" Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.386879 4968 scope.go:117] "RemoveContainer" containerID="11cab75071b70b8832e40c7683f03cc3798351958a69d001e69f6d4392e85adc" Feb 18 15:44:48 crc kubenswrapper[4968]: E0218 15:44:48.387290 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11cab75071b70b8832e40c7683f03cc3798351958a69d001e69f6d4392e85adc\": container with ID starting with 11cab75071b70b8832e40c7683f03cc3798351958a69d001e69f6d4392e85adc not found: ID does not exist" containerID="11cab75071b70b8832e40c7683f03cc3798351958a69d001e69f6d4392e85adc" Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.387467 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11cab75071b70b8832e40c7683f03cc3798351958a69d001e69f6d4392e85adc"} err="failed to get container status \"11cab75071b70b8832e40c7683f03cc3798351958a69d001e69f6d4392e85adc\": rpc error: code = NotFound desc = could not find container \"11cab75071b70b8832e40c7683f03cc3798351958a69d001e69f6d4392e85adc\": container with ID starting with 11cab75071b70b8832e40c7683f03cc3798351958a69d001e69f6d4392e85adc not found: ID does not exist" Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.387610 4968 scope.go:117] "RemoveContainer" containerID="1e31eb7ea8870715cf7f2e37b7cb29168240b9af6bd97361e9b258ce927a6e4e" Feb 18 15:44:48 crc kubenswrapper[4968]: E0218 15:44:48.388123 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e31eb7ea8870715cf7f2e37b7cb29168240b9af6bd97361e9b258ce927a6e4e\": container with ID starting with 1e31eb7ea8870715cf7f2e37b7cb29168240b9af6bd97361e9b258ce927a6e4e not found: ID does not exist" containerID="1e31eb7ea8870715cf7f2e37b7cb29168240b9af6bd97361e9b258ce927a6e4e" Feb 18 15:44:48 crc kubenswrapper[4968]: I0218 15:44:48.388169 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e31eb7ea8870715cf7f2e37b7cb29168240b9af6bd97361e9b258ce927a6e4e"} err="failed to get container status \"1e31eb7ea8870715cf7f2e37b7cb29168240b9af6bd97361e9b258ce927a6e4e\": rpc error: code = NotFound desc = could not find container \"1e31eb7ea8870715cf7f2e37b7cb29168240b9af6bd97361e9b258ce927a6e4e\": container with ID starting with 1e31eb7ea8870715cf7f2e37b7cb29168240b9af6bd97361e9b258ce927a6e4e not found: ID does not exist" Feb 18 15:44:49 crc kubenswrapper[4968]: I0218 15:44:49.241169 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a59d494a-c4b6-4673-a9a1-f85100b52ef0" path="/var/lib/kubelet/pods/a59d494a-c4b6-4673-a9a1-f85100b52ef0/volumes" Feb 18 15:44:54 crc kubenswrapper[4968]: I0218 15:44:54.114275 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:54 crc kubenswrapper[4968]: I0218 15:44:54.114682 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:54 crc kubenswrapper[4968]: I0218 15:44:54.161275 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:54 crc kubenswrapper[4968]: I0218 15:44:54.409140 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w2sfw" Feb 18 15:44:54 crc kubenswrapper[4968]: I0218 15:44:54.484133 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w2sfw"] Feb 18 15:44:54 crc kubenswrapper[4968]: I0218 15:44:54.522958 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9v9wz"] Feb 18 15:44:54 crc kubenswrapper[4968]: I0218 15:44:54.523194 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9v9wz" podUID="d3c68c1d-979b-4059-96bb-8c1415fb9776" containerName="registry-server" containerID="cri-o://0137e757cab2198962840a05a3de784bd956983ad77c50f73c21a64f3dce3a21" gracePeriod=2 Feb 18 15:44:55 crc kubenswrapper[4968]: I0218 15:44:55.370989 4968 generic.go:334] "Generic (PLEG): container finished" podID="d3c68c1d-979b-4059-96bb-8c1415fb9776" containerID="0137e757cab2198962840a05a3de784bd956983ad77c50f73c21a64f3dce3a21" exitCode=0 Feb 18 15:44:55 crc kubenswrapper[4968]: I0218 15:44:55.371077 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9v9wz" event={"ID":"d3c68c1d-979b-4059-96bb-8c1415fb9776","Type":"ContainerDied","Data":"0137e757cab2198962840a05a3de784bd956983ad77c50f73c21a64f3dce3a21"} Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.026589 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.120536 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3c68c1d-979b-4059-96bb-8c1415fb9776-utilities\") pod \"d3c68c1d-979b-4059-96bb-8c1415fb9776\" (UID: \"d3c68c1d-979b-4059-96bb-8c1415fb9776\") " Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.120670 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xth4w\" (UniqueName: \"kubernetes.io/projected/d3c68c1d-979b-4059-96bb-8c1415fb9776-kube-api-access-xth4w\") pod \"d3c68c1d-979b-4059-96bb-8c1415fb9776\" (UID: \"d3c68c1d-979b-4059-96bb-8c1415fb9776\") " Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.120793 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3c68c1d-979b-4059-96bb-8c1415fb9776-catalog-content\") pod \"d3c68c1d-979b-4059-96bb-8c1415fb9776\" (UID: \"d3c68c1d-979b-4059-96bb-8c1415fb9776\") " Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.121293 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3c68c1d-979b-4059-96bb-8c1415fb9776-utilities" (OuterVolumeSpecName: "utilities") pod "d3c68c1d-979b-4059-96bb-8c1415fb9776" (UID: "d3c68c1d-979b-4059-96bb-8c1415fb9776"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.126976 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3c68c1d-979b-4059-96bb-8c1415fb9776-kube-api-access-xth4w" (OuterVolumeSpecName: "kube-api-access-xth4w") pod "d3c68c1d-979b-4059-96bb-8c1415fb9776" (UID: "d3c68c1d-979b-4059-96bb-8c1415fb9776"). InnerVolumeSpecName "kube-api-access-xth4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.221834 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3c68c1d-979b-4059-96bb-8c1415fb9776-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.221865 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xth4w\" (UniqueName: \"kubernetes.io/projected/d3c68c1d-979b-4059-96bb-8c1415fb9776-kube-api-access-xth4w\") on node \"crc\" DevicePath \"\"" Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.244296 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3c68c1d-979b-4059-96bb-8c1415fb9776-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3c68c1d-979b-4059-96bb-8c1415fb9776" (UID: "d3c68c1d-979b-4059-96bb-8c1415fb9776"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.322881 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3c68c1d-979b-4059-96bb-8c1415fb9776-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.379967 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9v9wz" event={"ID":"d3c68c1d-979b-4059-96bb-8c1415fb9776","Type":"ContainerDied","Data":"f9e21aa7dbf46e9e6047a5a8840d655228831acf92393bc58b2179698d93b214"} Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.380025 4968 scope.go:117] "RemoveContainer" containerID="0137e757cab2198962840a05a3de784bd956983ad77c50f73c21a64f3dce3a21" Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.379986 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9v9wz" Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.400324 4968 scope.go:117] "RemoveContainer" containerID="179574daa388c576258add2f3fc3c6757325e019284e9f1e232e82813a558272" Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.408930 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9v9wz"] Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.438008 4968 scope.go:117] "RemoveContainer" containerID="eeb56d4fee21f21ea4d0bd26231a6abff321639e86438c99b46ab948265b489b" Feb 18 15:44:56 crc kubenswrapper[4968]: I0218 15:44:56.441435 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9v9wz"] Feb 18 15:44:57 crc kubenswrapper[4968]: I0218 15:44:57.240192 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3c68c1d-979b-4059-96bb-8c1415fb9776" path="/var/lib/kubelet/pods/d3c68c1d-979b-4059-96bb-8c1415fb9776/volumes" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.142016 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm"] Feb 18 15:45:00 crc kubenswrapper[4968]: E0218 15:45:00.142644 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3c68c1d-979b-4059-96bb-8c1415fb9776" containerName="extract-utilities" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.142660 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3c68c1d-979b-4059-96bb-8c1415fb9776" containerName="extract-utilities" Feb 18 15:45:00 crc kubenswrapper[4968]: E0218 15:45:00.142677 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a59d494a-c4b6-4673-a9a1-f85100b52ef0" containerName="extract-content" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.142685 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="a59d494a-c4b6-4673-a9a1-f85100b52ef0" containerName="extract-content" Feb 18 15:45:00 crc kubenswrapper[4968]: E0218 15:45:00.142705 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3c68c1d-979b-4059-96bb-8c1415fb9776" containerName="extract-content" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.142713 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3c68c1d-979b-4059-96bb-8c1415fb9776" containerName="extract-content" Feb 18 15:45:00 crc kubenswrapper[4968]: E0218 15:45:00.142721 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a59d494a-c4b6-4673-a9a1-f85100b52ef0" containerName="extract-utilities" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.142728 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="a59d494a-c4b6-4673-a9a1-f85100b52ef0" containerName="extract-utilities" Feb 18 15:45:00 crc kubenswrapper[4968]: E0218 15:45:00.142741 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3c68c1d-979b-4059-96bb-8c1415fb9776" containerName="registry-server" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.142765 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3c68c1d-979b-4059-96bb-8c1415fb9776" containerName="registry-server" Feb 18 15:45:00 crc kubenswrapper[4968]: E0218 15:45:00.142777 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a59d494a-c4b6-4673-a9a1-f85100b52ef0" containerName="registry-server" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.142783 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="a59d494a-c4b6-4673-a9a1-f85100b52ef0" containerName="registry-server" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.142974 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3c68c1d-979b-4059-96bb-8c1415fb9776" containerName="registry-server" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.143001 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="a59d494a-c4b6-4673-a9a1-f85100b52ef0" containerName="registry-server" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.143559 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.148723 4968 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.148824 4968 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.151884 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm"] Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.277043 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-secret-volume\") pod \"collect-profiles-29523825-dm9dm\" (UID: \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.277095 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf9t5\" (UniqueName: \"kubernetes.io/projected/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-kube-api-access-nf9t5\") pod \"collect-profiles-29523825-dm9dm\" (UID: \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.277194 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-config-volume\") pod \"collect-profiles-29523825-dm9dm\" (UID: \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.378402 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-config-volume\") pod \"collect-profiles-29523825-dm9dm\" (UID: \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.378499 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-secret-volume\") pod \"collect-profiles-29523825-dm9dm\" (UID: \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.378529 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf9t5\" (UniqueName: \"kubernetes.io/projected/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-kube-api-access-nf9t5\") pod \"collect-profiles-29523825-dm9dm\" (UID: \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.379588 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-config-volume\") pod \"collect-profiles-29523825-dm9dm\" (UID: \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.383838 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-secret-volume\") pod \"collect-profiles-29523825-dm9dm\" (UID: \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.400956 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf9t5\" (UniqueName: \"kubernetes.io/projected/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-kube-api-access-nf9t5\") pod \"collect-profiles-29523825-dm9dm\" (UID: \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.461979 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:00 crc kubenswrapper[4968]: I0218 15:45:00.872258 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm"] Feb 18 15:45:01 crc kubenswrapper[4968]: I0218 15:45:01.424320 4968 generic.go:334] "Generic (PLEG): container finished" podID="d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92" containerID="dccbbc2188a4686675fd0e4b7285839408c78f1e8a8f89f5ac80f43760b03bf7" exitCode=0 Feb 18 15:45:01 crc kubenswrapper[4968]: I0218 15:45:01.424376 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" event={"ID":"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92","Type":"ContainerDied","Data":"dccbbc2188a4686675fd0e4b7285839408c78f1e8a8f89f5ac80f43760b03bf7"} Feb 18 15:45:01 crc kubenswrapper[4968]: I0218 15:45:01.424420 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" event={"ID":"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92","Type":"ContainerStarted","Data":"e1ce879c24c950e32e0fb3de86bed1fb5d2ffe2c273cadde84ffa05758530cb1"} Feb 18 15:45:02 crc kubenswrapper[4968]: I0218 15:45:02.694856 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:02 crc kubenswrapper[4968]: I0218 15:45:02.811400 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-secret-volume\") pod \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\" (UID: \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\") " Feb 18 15:45:02 crc kubenswrapper[4968]: I0218 15:45:02.811556 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf9t5\" (UniqueName: \"kubernetes.io/projected/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-kube-api-access-nf9t5\") pod \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\" (UID: \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\") " Feb 18 15:45:02 crc kubenswrapper[4968]: I0218 15:45:02.811590 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-config-volume\") pod \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\" (UID: \"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92\") " Feb 18 15:45:02 crc kubenswrapper[4968]: I0218 15:45:02.812516 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-config-volume" (OuterVolumeSpecName: "config-volume") pod "d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92" (UID: "d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:45:02 crc kubenswrapper[4968]: I0218 15:45:02.816924 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-kube-api-access-nf9t5" (OuterVolumeSpecName: "kube-api-access-nf9t5") pod "d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92" (UID: "d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92"). InnerVolumeSpecName "kube-api-access-nf9t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:45:02 crc kubenswrapper[4968]: I0218 15:45:02.818587 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92" (UID: "d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:45:02 crc kubenswrapper[4968]: I0218 15:45:02.913227 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf9t5\" (UniqueName: \"kubernetes.io/projected/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-kube-api-access-nf9t5\") on node \"crc\" DevicePath \"\"" Feb 18 15:45:02 crc kubenswrapper[4968]: I0218 15:45:02.913279 4968 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 15:45:02 crc kubenswrapper[4968]: I0218 15:45:02.913298 4968 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 18 15:45:03 crc kubenswrapper[4968]: I0218 15:45:03.439983 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" event={"ID":"d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92","Type":"ContainerDied","Data":"e1ce879c24c950e32e0fb3de86bed1fb5d2ffe2c273cadde84ffa05758530cb1"} Feb 18 15:45:03 crc kubenswrapper[4968]: I0218 15:45:03.440021 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523825-dm9dm" Feb 18 15:45:03 crc kubenswrapper[4968]: I0218 15:45:03.440023 4968 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1ce879c24c950e32e0fb3de86bed1fb5d2ffe2c273cadde84ffa05758530cb1" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.337183 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g5mrs"] Feb 18 15:46:04 crc kubenswrapper[4968]: E0218 15:46:04.338130 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92" containerName="collect-profiles" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.338144 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92" containerName="collect-profiles" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.338305 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6b3b51d-f47a-4a2f-afaf-c07ecdc51c92" containerName="collect-profiles" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.339337 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.350999 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g5mrs"] Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.535735 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7921d576-1aea-4644-b415-8a30eea6c682-utilities\") pod \"certified-operators-g5mrs\" (UID: \"7921d576-1aea-4644-b415-8a30eea6c682\") " pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.536145 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7921d576-1aea-4644-b415-8a30eea6c682-catalog-content\") pod \"certified-operators-g5mrs\" (UID: \"7921d576-1aea-4644-b415-8a30eea6c682\") " pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.536176 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c5rg\" (UniqueName: \"kubernetes.io/projected/7921d576-1aea-4644-b415-8a30eea6c682-kube-api-access-7c5rg\") pod \"certified-operators-g5mrs\" (UID: \"7921d576-1aea-4644-b415-8a30eea6c682\") " pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.636941 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7921d576-1aea-4644-b415-8a30eea6c682-utilities\") pod \"certified-operators-g5mrs\" (UID: \"7921d576-1aea-4644-b415-8a30eea6c682\") " pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.637002 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7921d576-1aea-4644-b415-8a30eea6c682-catalog-content\") pod \"certified-operators-g5mrs\" (UID: \"7921d576-1aea-4644-b415-8a30eea6c682\") " pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.637028 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c5rg\" (UniqueName: \"kubernetes.io/projected/7921d576-1aea-4644-b415-8a30eea6c682-kube-api-access-7c5rg\") pod \"certified-operators-g5mrs\" (UID: \"7921d576-1aea-4644-b415-8a30eea6c682\") " pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.637678 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7921d576-1aea-4644-b415-8a30eea6c682-catalog-content\") pod \"certified-operators-g5mrs\" (UID: \"7921d576-1aea-4644-b415-8a30eea6c682\") " pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.637893 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7921d576-1aea-4644-b415-8a30eea6c682-utilities\") pod \"certified-operators-g5mrs\" (UID: \"7921d576-1aea-4644-b415-8a30eea6c682\") " pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.664257 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c5rg\" (UniqueName: \"kubernetes.io/projected/7921d576-1aea-4644-b415-8a30eea6c682-kube-api-access-7c5rg\") pod \"certified-operators-g5mrs\" (UID: \"7921d576-1aea-4644-b415-8a30eea6c682\") " pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.671153 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:04 crc kubenswrapper[4968]: I0218 15:46:04.922831 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g5mrs"] Feb 18 15:46:05 crc kubenswrapper[4968]: I0218 15:46:05.910115 4968 generic.go:334] "Generic (PLEG): container finished" podID="7921d576-1aea-4644-b415-8a30eea6c682" containerID="62afe35448d51f75ab194abcca5ac65a40e86f434280881670c075ae55aa34f1" exitCode=0 Feb 18 15:46:05 crc kubenswrapper[4968]: I0218 15:46:05.910161 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5mrs" event={"ID":"7921d576-1aea-4644-b415-8a30eea6c682","Type":"ContainerDied","Data":"62afe35448d51f75ab194abcca5ac65a40e86f434280881670c075ae55aa34f1"} Feb 18 15:46:05 crc kubenswrapper[4968]: I0218 15:46:05.910435 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5mrs" event={"ID":"7921d576-1aea-4644-b415-8a30eea6c682","Type":"ContainerStarted","Data":"5b830bb441a4a914b7ccfa453a54892da379520f59cdb844d348c7c2e591a10f"} Feb 18 15:46:07 crc kubenswrapper[4968]: I0218 15:46:07.929614 4968 generic.go:334] "Generic (PLEG): container finished" podID="7921d576-1aea-4644-b415-8a30eea6c682" containerID="a30bff409f6058867c81c6f13c79e60b7e57cea74217d5e6be6b6c9cfff9cb17" exitCode=0 Feb 18 15:46:07 crc kubenswrapper[4968]: I0218 15:46:07.929668 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5mrs" event={"ID":"7921d576-1aea-4644-b415-8a30eea6c682","Type":"ContainerDied","Data":"a30bff409f6058867c81c6f13c79e60b7e57cea74217d5e6be6b6c9cfff9cb17"} Feb 18 15:46:10 crc kubenswrapper[4968]: I0218 15:46:10.953925 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5mrs" event={"ID":"7921d576-1aea-4644-b415-8a30eea6c682","Type":"ContainerStarted","Data":"2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5"} Feb 18 15:46:10 crc kubenswrapper[4968]: I0218 15:46:10.984161 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g5mrs" podStartSLOduration=2.923793818 podStartE2EDuration="6.984142212s" podCreationTimestamp="2026-02-18 15:46:04 +0000 UTC" firstStartedPulling="2026-02-18 15:46:05.913048089 +0000 UTC m=+1465.298492951" lastFinishedPulling="2026-02-18 15:46:09.973396483 +0000 UTC m=+1469.358841345" observedRunningTime="2026-02-18 15:46:10.977978328 +0000 UTC m=+1470.363423200" watchObservedRunningTime="2026-02-18 15:46:10.984142212 +0000 UTC m=+1470.369587074" Feb 18 15:46:14 crc kubenswrapper[4968]: I0218 15:46:14.367805 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:46:14 crc kubenswrapper[4968]: I0218 15:46:14.368318 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:46:14 crc kubenswrapper[4968]: I0218 15:46:14.672142 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:14 crc kubenswrapper[4968]: I0218 15:46:14.672466 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:14 crc kubenswrapper[4968]: I0218 15:46:14.742024 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:15 crc kubenswrapper[4968]: I0218 15:46:15.106584 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:15 crc kubenswrapper[4968]: I0218 15:46:15.169711 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g5mrs"] Feb 18 15:46:17 crc kubenswrapper[4968]: I0218 15:46:17.049739 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g5mrs" podUID="7921d576-1aea-4644-b415-8a30eea6c682" containerName="registry-server" containerID="cri-o://2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5" gracePeriod=2 Feb 18 15:46:17 crc kubenswrapper[4968]: I0218 15:46:17.423447 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:17 crc kubenswrapper[4968]: I0218 15:46:17.576338 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7921d576-1aea-4644-b415-8a30eea6c682-utilities\") pod \"7921d576-1aea-4644-b415-8a30eea6c682\" (UID: \"7921d576-1aea-4644-b415-8a30eea6c682\") " Feb 18 15:46:17 crc kubenswrapper[4968]: I0218 15:46:17.576494 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7921d576-1aea-4644-b415-8a30eea6c682-catalog-content\") pod \"7921d576-1aea-4644-b415-8a30eea6c682\" (UID: \"7921d576-1aea-4644-b415-8a30eea6c682\") " Feb 18 15:46:17 crc kubenswrapper[4968]: I0218 15:46:17.576545 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c5rg\" (UniqueName: \"kubernetes.io/projected/7921d576-1aea-4644-b415-8a30eea6c682-kube-api-access-7c5rg\") pod \"7921d576-1aea-4644-b415-8a30eea6c682\" (UID: \"7921d576-1aea-4644-b415-8a30eea6c682\") " Feb 18 15:46:17 crc kubenswrapper[4968]: I0218 15:46:17.577291 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7921d576-1aea-4644-b415-8a30eea6c682-utilities" (OuterVolumeSpecName: "utilities") pod "7921d576-1aea-4644-b415-8a30eea6c682" (UID: "7921d576-1aea-4644-b415-8a30eea6c682"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:46:17 crc kubenswrapper[4968]: I0218 15:46:17.581727 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7921d576-1aea-4644-b415-8a30eea6c682-kube-api-access-7c5rg" (OuterVolumeSpecName: "kube-api-access-7c5rg") pod "7921d576-1aea-4644-b415-8a30eea6c682" (UID: "7921d576-1aea-4644-b415-8a30eea6c682"). InnerVolumeSpecName "kube-api-access-7c5rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:46:17 crc kubenswrapper[4968]: I0218 15:46:17.678684 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c5rg\" (UniqueName: \"kubernetes.io/projected/7921d576-1aea-4644-b415-8a30eea6c682-kube-api-access-7c5rg\") on node \"crc\" DevicePath \"\"" Feb 18 15:46:17 crc kubenswrapper[4968]: I0218 15:46:17.678717 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7921d576-1aea-4644-b415-8a30eea6c682-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.063109 4968 generic.go:334] "Generic (PLEG): container finished" podID="7921d576-1aea-4644-b415-8a30eea6c682" containerID="2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5" exitCode=0 Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.063179 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5mrs" event={"ID":"7921d576-1aea-4644-b415-8a30eea6c682","Type":"ContainerDied","Data":"2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5"} Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.063228 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5mrs" event={"ID":"7921d576-1aea-4644-b415-8a30eea6c682","Type":"ContainerDied","Data":"5b830bb441a4a914b7ccfa453a54892da379520f59cdb844d348c7c2e591a10f"} Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.063292 4968 scope.go:117] "RemoveContainer" containerID="2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.063486 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5mrs" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.105660 4968 scope.go:117] "RemoveContainer" containerID="a30bff409f6058867c81c6f13c79e60b7e57cea74217d5e6be6b6c9cfff9cb17" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.125874 4968 scope.go:117] "RemoveContainer" containerID="62afe35448d51f75ab194abcca5ac65a40e86f434280881670c075ae55aa34f1" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.154626 4968 scope.go:117] "RemoveContainer" containerID="2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5" Feb 18 15:46:18 crc kubenswrapper[4968]: E0218 15:46:18.154991 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5\": container with ID starting with 2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5 not found: ID does not exist" containerID="2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.155034 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5"} err="failed to get container status \"2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5\": rpc error: code = NotFound desc = could not find container \"2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5\": container with ID starting with 2f42e73eac6043453130ae311f3fa30cf26424241a26a2200d5490bee63c89e5 not found: ID does not exist" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.155062 4968 scope.go:117] "RemoveContainer" containerID="a30bff409f6058867c81c6f13c79e60b7e57cea74217d5e6be6b6c9cfff9cb17" Feb 18 15:46:18 crc kubenswrapper[4968]: E0218 15:46:18.155448 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a30bff409f6058867c81c6f13c79e60b7e57cea74217d5e6be6b6c9cfff9cb17\": container with ID starting with a30bff409f6058867c81c6f13c79e60b7e57cea74217d5e6be6b6c9cfff9cb17 not found: ID does not exist" containerID="a30bff409f6058867c81c6f13c79e60b7e57cea74217d5e6be6b6c9cfff9cb17" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.155486 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a30bff409f6058867c81c6f13c79e60b7e57cea74217d5e6be6b6c9cfff9cb17"} err="failed to get container status \"a30bff409f6058867c81c6f13c79e60b7e57cea74217d5e6be6b6c9cfff9cb17\": rpc error: code = NotFound desc = could not find container \"a30bff409f6058867c81c6f13c79e60b7e57cea74217d5e6be6b6c9cfff9cb17\": container with ID starting with a30bff409f6058867c81c6f13c79e60b7e57cea74217d5e6be6b6c9cfff9cb17 not found: ID does not exist" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.155507 4968 scope.go:117] "RemoveContainer" containerID="62afe35448d51f75ab194abcca5ac65a40e86f434280881670c075ae55aa34f1" Feb 18 15:46:18 crc kubenswrapper[4968]: E0218 15:46:18.155811 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62afe35448d51f75ab194abcca5ac65a40e86f434280881670c075ae55aa34f1\": container with ID starting with 62afe35448d51f75ab194abcca5ac65a40e86f434280881670c075ae55aa34f1 not found: ID does not exist" containerID="62afe35448d51f75ab194abcca5ac65a40e86f434280881670c075ae55aa34f1" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.155933 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62afe35448d51f75ab194abcca5ac65a40e86f434280881670c075ae55aa34f1"} err="failed to get container status \"62afe35448d51f75ab194abcca5ac65a40e86f434280881670c075ae55aa34f1\": rpc error: code = NotFound desc = could not find container \"62afe35448d51f75ab194abcca5ac65a40e86f434280881670c075ae55aa34f1\": container with ID starting with 62afe35448d51f75ab194abcca5ac65a40e86f434280881670c075ae55aa34f1 not found: ID does not exist" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.224504 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7921d576-1aea-4644-b415-8a30eea6c682-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7921d576-1aea-4644-b415-8a30eea6c682" (UID: "7921d576-1aea-4644-b415-8a30eea6c682"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.289977 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7921d576-1aea-4644-b415-8a30eea6c682-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.395926 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g5mrs"] Feb 18 15:46:18 crc kubenswrapper[4968]: I0218 15:46:18.402419 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g5mrs"] Feb 18 15:46:19 crc kubenswrapper[4968]: I0218 15:46:19.249179 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7921d576-1aea-4644-b415-8a30eea6c682" path="/var/lib/kubelet/pods/7921d576-1aea-4644-b415-8a30eea6c682/volumes" Feb 18 15:46:44 crc kubenswrapper[4968]: I0218 15:46:44.427971 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:46:44 crc kubenswrapper[4968]: I0218 15:46:44.428598 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.173096 4968 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cctfk"] Feb 18 15:46:45 crc kubenswrapper[4968]: E0218 15:46:45.173799 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7921d576-1aea-4644-b415-8a30eea6c682" containerName="extract-utilities" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.173817 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="7921d576-1aea-4644-b415-8a30eea6c682" containerName="extract-utilities" Feb 18 15:46:45 crc kubenswrapper[4968]: E0218 15:46:45.173840 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7921d576-1aea-4644-b415-8a30eea6c682" containerName="registry-server" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.173847 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="7921d576-1aea-4644-b415-8a30eea6c682" containerName="registry-server" Feb 18 15:46:45 crc kubenswrapper[4968]: E0218 15:46:45.173874 4968 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7921d576-1aea-4644-b415-8a30eea6c682" containerName="extract-content" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.173883 4968 state_mem.go:107] "Deleted CPUSet assignment" podUID="7921d576-1aea-4644-b415-8a30eea6c682" containerName="extract-content" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.174035 4968 memory_manager.go:354] "RemoveStaleState removing state" podUID="7921d576-1aea-4644-b415-8a30eea6c682" containerName="registry-server" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.175483 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.180640 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cctfk"] Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.340556 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a76405d-fc84-4802-a845-b6f6dfdc2552-utilities\") pod \"community-operators-cctfk\" (UID: \"6a76405d-fc84-4802-a845-b6f6dfdc2552\") " pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.341503 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwlqz\" (UniqueName: \"kubernetes.io/projected/6a76405d-fc84-4802-a845-b6f6dfdc2552-kube-api-access-fwlqz\") pod \"community-operators-cctfk\" (UID: \"6a76405d-fc84-4802-a845-b6f6dfdc2552\") " pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.341624 4968 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a76405d-fc84-4802-a845-b6f6dfdc2552-catalog-content\") pod \"community-operators-cctfk\" (UID: \"6a76405d-fc84-4802-a845-b6f6dfdc2552\") " pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.442508 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a76405d-fc84-4802-a845-b6f6dfdc2552-utilities\") pod \"community-operators-cctfk\" (UID: \"6a76405d-fc84-4802-a845-b6f6dfdc2552\") " pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.442549 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwlqz\" (UniqueName: \"kubernetes.io/projected/6a76405d-fc84-4802-a845-b6f6dfdc2552-kube-api-access-fwlqz\") pod \"community-operators-cctfk\" (UID: \"6a76405d-fc84-4802-a845-b6f6dfdc2552\") " pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.442589 4968 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a76405d-fc84-4802-a845-b6f6dfdc2552-catalog-content\") pod \"community-operators-cctfk\" (UID: \"6a76405d-fc84-4802-a845-b6f6dfdc2552\") " pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.442988 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a76405d-fc84-4802-a845-b6f6dfdc2552-utilities\") pod \"community-operators-cctfk\" (UID: \"6a76405d-fc84-4802-a845-b6f6dfdc2552\") " pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.443020 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a76405d-fc84-4802-a845-b6f6dfdc2552-catalog-content\") pod \"community-operators-cctfk\" (UID: \"6a76405d-fc84-4802-a845-b6f6dfdc2552\") " pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.475240 4968 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwlqz\" (UniqueName: \"kubernetes.io/projected/6a76405d-fc84-4802-a845-b6f6dfdc2552-kube-api-access-fwlqz\") pod \"community-operators-cctfk\" (UID: \"6a76405d-fc84-4802-a845-b6f6dfdc2552\") " pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.496147 4968 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:45 crc kubenswrapper[4968]: I0218 15:46:45.981562 4968 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cctfk"] Feb 18 15:46:46 crc kubenswrapper[4968]: I0218 15:46:46.452288 4968 generic.go:334] "Generic (PLEG): container finished" podID="6a76405d-fc84-4802-a845-b6f6dfdc2552" containerID="f4a07930214863168c931dcbec59edb8776cb31784b9cd530026572a4c5a595c" exitCode=0 Feb 18 15:46:46 crc kubenswrapper[4968]: I0218 15:46:46.452389 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cctfk" event={"ID":"6a76405d-fc84-4802-a845-b6f6dfdc2552","Type":"ContainerDied","Data":"f4a07930214863168c931dcbec59edb8776cb31784b9cd530026572a4c5a595c"} Feb 18 15:46:46 crc kubenswrapper[4968]: I0218 15:46:46.452634 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cctfk" event={"ID":"6a76405d-fc84-4802-a845-b6f6dfdc2552","Type":"ContainerStarted","Data":"3086d71d1fc811d534e803a14ef79c87d38373f6ea5646a58d9fed757c19a351"} Feb 18 15:46:47 crc kubenswrapper[4968]: I0218 15:46:47.461647 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cctfk" event={"ID":"6a76405d-fc84-4802-a845-b6f6dfdc2552","Type":"ContainerStarted","Data":"37c06dceaf04958a5e9893b1bbba7f4856970c6466b5753b1aa44a0cd1a03b83"} Feb 18 15:46:48 crc kubenswrapper[4968]: I0218 15:46:48.470561 4968 generic.go:334] "Generic (PLEG): container finished" podID="6a76405d-fc84-4802-a845-b6f6dfdc2552" containerID="37c06dceaf04958a5e9893b1bbba7f4856970c6466b5753b1aa44a0cd1a03b83" exitCode=0 Feb 18 15:46:48 crc kubenswrapper[4968]: I0218 15:46:48.470613 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cctfk" event={"ID":"6a76405d-fc84-4802-a845-b6f6dfdc2552","Type":"ContainerDied","Data":"37c06dceaf04958a5e9893b1bbba7f4856970c6466b5753b1aa44a0cd1a03b83"} Feb 18 15:46:49 crc kubenswrapper[4968]: I0218 15:46:49.480344 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cctfk" event={"ID":"6a76405d-fc84-4802-a845-b6f6dfdc2552","Type":"ContainerStarted","Data":"4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd"} Feb 18 15:46:49 crc kubenswrapper[4968]: I0218 15:46:49.511769 4968 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cctfk" podStartSLOduration=1.735599788 podStartE2EDuration="4.511728402s" podCreationTimestamp="2026-02-18 15:46:45 +0000 UTC" firstStartedPulling="2026-02-18 15:46:46.454034418 +0000 UTC m=+1505.839479280" lastFinishedPulling="2026-02-18 15:46:49.230163032 +0000 UTC m=+1508.615607894" observedRunningTime="2026-02-18 15:46:49.506017651 +0000 UTC m=+1508.891462513" watchObservedRunningTime="2026-02-18 15:46:49.511728402 +0000 UTC m=+1508.897173274" Feb 18 15:46:55 crc kubenswrapper[4968]: I0218 15:46:55.496576 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:55 crc kubenswrapper[4968]: I0218 15:46:55.497062 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:55 crc kubenswrapper[4968]: I0218 15:46:55.546096 4968 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:55 crc kubenswrapper[4968]: I0218 15:46:55.597918 4968 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:55 crc kubenswrapper[4968]: I0218 15:46:55.786459 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cctfk"] Feb 18 15:46:57 crc kubenswrapper[4968]: I0218 15:46:57.548862 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cctfk" podUID="6a76405d-fc84-4802-a845-b6f6dfdc2552" containerName="registry-server" containerID="cri-o://4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd" gracePeriod=2 Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.192314 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.341023 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a76405d-fc84-4802-a845-b6f6dfdc2552-catalog-content\") pod \"6a76405d-fc84-4802-a845-b6f6dfdc2552\" (UID: \"6a76405d-fc84-4802-a845-b6f6dfdc2552\") " Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.341102 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwlqz\" (UniqueName: \"kubernetes.io/projected/6a76405d-fc84-4802-a845-b6f6dfdc2552-kube-api-access-fwlqz\") pod \"6a76405d-fc84-4802-a845-b6f6dfdc2552\" (UID: \"6a76405d-fc84-4802-a845-b6f6dfdc2552\") " Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.341137 4968 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a76405d-fc84-4802-a845-b6f6dfdc2552-utilities\") pod \"6a76405d-fc84-4802-a845-b6f6dfdc2552\" (UID: \"6a76405d-fc84-4802-a845-b6f6dfdc2552\") " Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.341978 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a76405d-fc84-4802-a845-b6f6dfdc2552-utilities" (OuterVolumeSpecName: "utilities") pod "6a76405d-fc84-4802-a845-b6f6dfdc2552" (UID: "6a76405d-fc84-4802-a845-b6f6dfdc2552"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.347023 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a76405d-fc84-4802-a845-b6f6dfdc2552-kube-api-access-fwlqz" (OuterVolumeSpecName: "kube-api-access-fwlqz") pod "6a76405d-fc84-4802-a845-b6f6dfdc2552" (UID: "6a76405d-fc84-4802-a845-b6f6dfdc2552"). InnerVolumeSpecName "kube-api-access-fwlqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.442329 4968 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwlqz\" (UniqueName: \"kubernetes.io/projected/6a76405d-fc84-4802-a845-b6f6dfdc2552-kube-api-access-fwlqz\") on node \"crc\" DevicePath \"\"" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.442363 4968 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a76405d-fc84-4802-a845-b6f6dfdc2552-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.557849 4968 generic.go:334] "Generic (PLEG): container finished" podID="6a76405d-fc84-4802-a845-b6f6dfdc2552" containerID="4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd" exitCode=0 Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.557890 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cctfk" event={"ID":"6a76405d-fc84-4802-a845-b6f6dfdc2552","Type":"ContainerDied","Data":"4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd"} Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.557916 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cctfk" event={"ID":"6a76405d-fc84-4802-a845-b6f6dfdc2552","Type":"ContainerDied","Data":"3086d71d1fc811d534e803a14ef79c87d38373f6ea5646a58d9fed757c19a351"} Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.557932 4968 scope.go:117] "RemoveContainer" containerID="4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.559054 4968 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cctfk" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.575360 4968 scope.go:117] "RemoveContainer" containerID="37c06dceaf04958a5e9893b1bbba7f4856970c6466b5753b1aa44a0cd1a03b83" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.585267 4968 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a76405d-fc84-4802-a845-b6f6dfdc2552-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a76405d-fc84-4802-a845-b6f6dfdc2552" (UID: "6a76405d-fc84-4802-a845-b6f6dfdc2552"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.594955 4968 scope.go:117] "RemoveContainer" containerID="f4a07930214863168c931dcbec59edb8776cb31784b9cd530026572a4c5a595c" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.626085 4968 scope.go:117] "RemoveContainer" containerID="4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd" Feb 18 15:46:58 crc kubenswrapper[4968]: E0218 15:46:58.626735 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd\": container with ID starting with 4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd not found: ID does not exist" containerID="4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.626796 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd"} err="failed to get container status \"4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd\": rpc error: code = NotFound desc = could not find container \"4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd\": container with ID starting with 4aec922dd530096d5436d49e215d06902a799939607f72cbde59ecbc9064e5dd not found: ID does not exist" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.626824 4968 scope.go:117] "RemoveContainer" containerID="37c06dceaf04958a5e9893b1bbba7f4856970c6466b5753b1aa44a0cd1a03b83" Feb 18 15:46:58 crc kubenswrapper[4968]: E0218 15:46:58.627251 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37c06dceaf04958a5e9893b1bbba7f4856970c6466b5753b1aa44a0cd1a03b83\": container with ID starting with 37c06dceaf04958a5e9893b1bbba7f4856970c6466b5753b1aa44a0cd1a03b83 not found: ID does not exist" containerID="37c06dceaf04958a5e9893b1bbba7f4856970c6466b5753b1aa44a0cd1a03b83" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.627289 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37c06dceaf04958a5e9893b1bbba7f4856970c6466b5753b1aa44a0cd1a03b83"} err="failed to get container status \"37c06dceaf04958a5e9893b1bbba7f4856970c6466b5753b1aa44a0cd1a03b83\": rpc error: code = NotFound desc = could not find container \"37c06dceaf04958a5e9893b1bbba7f4856970c6466b5753b1aa44a0cd1a03b83\": container with ID starting with 37c06dceaf04958a5e9893b1bbba7f4856970c6466b5753b1aa44a0cd1a03b83 not found: ID does not exist" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.627341 4968 scope.go:117] "RemoveContainer" containerID="f4a07930214863168c931dcbec59edb8776cb31784b9cd530026572a4c5a595c" Feb 18 15:46:58 crc kubenswrapper[4968]: E0218 15:46:58.627706 4968 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4a07930214863168c931dcbec59edb8776cb31784b9cd530026572a4c5a595c\": container with ID starting with f4a07930214863168c931dcbec59edb8776cb31784b9cd530026572a4c5a595c not found: ID does not exist" containerID="f4a07930214863168c931dcbec59edb8776cb31784b9cd530026572a4c5a595c" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.627740 4968 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4a07930214863168c931dcbec59edb8776cb31784b9cd530026572a4c5a595c"} err="failed to get container status \"f4a07930214863168c931dcbec59edb8776cb31784b9cd530026572a4c5a595c\": rpc error: code = NotFound desc = could not find container \"f4a07930214863168c931dcbec59edb8776cb31784b9cd530026572a4c5a595c\": container with ID starting with f4a07930214863168c931dcbec59edb8776cb31784b9cd530026572a4c5a595c not found: ID does not exist" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.644912 4968 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a76405d-fc84-4802-a845-b6f6dfdc2552-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.897840 4968 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cctfk"] Feb 18 15:46:58 crc kubenswrapper[4968]: I0218 15:46:58.903116 4968 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cctfk"] Feb 18 15:46:59 crc kubenswrapper[4968]: I0218 15:46:59.239992 4968 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a76405d-fc84-4802-a845-b6f6dfdc2552" path="/var/lib/kubelet/pods/6a76405d-fc84-4802-a845-b6f6dfdc2552/volumes" Feb 18 15:47:14 crc kubenswrapper[4968]: I0218 15:47:14.367513 4968 patch_prober.go:28] interesting pod/machine-config-daemon-xnhwb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:47:14 crc kubenswrapper[4968]: I0218 15:47:14.368091 4968 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:47:14 crc kubenswrapper[4968]: I0218 15:47:14.368136 4968 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" Feb 18 15:47:14 crc kubenswrapper[4968]: I0218 15:47:14.368681 4968 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a5f596668788929acd43efc0faf07605f488d17a5ceabc4b158d494ee696c99c"} pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:47:14 crc kubenswrapper[4968]: I0218 15:47:14.368730 4968 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerName="machine-config-daemon" containerID="cri-o://a5f596668788929acd43efc0faf07605f488d17a5ceabc4b158d494ee696c99c" gracePeriod=600 Feb 18 15:47:14 crc kubenswrapper[4968]: E0218 15:47:14.491876 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xnhwb_openshift-machine-config-operator(f9bae90c-908f-40fd-8373-4bf7f9aaede6)\"" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" Feb 18 15:47:14 crc kubenswrapper[4968]: I0218 15:47:14.664993 4968 generic.go:334] "Generic (PLEG): container finished" podID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" containerID="a5f596668788929acd43efc0faf07605f488d17a5ceabc4b158d494ee696c99c" exitCode=0 Feb 18 15:47:14 crc kubenswrapper[4968]: I0218 15:47:14.665043 4968 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" event={"ID":"f9bae90c-908f-40fd-8373-4bf7f9aaede6","Type":"ContainerDied","Data":"a5f596668788929acd43efc0faf07605f488d17a5ceabc4b158d494ee696c99c"} Feb 18 15:47:14 crc kubenswrapper[4968]: I0218 15:47:14.665080 4968 scope.go:117] "RemoveContainer" containerID="79a4ac6e0d12350981bce67b06b3c00e9895b13a7171aa9010c1bcf9c6a8723e" Feb 18 15:47:14 crc kubenswrapper[4968]: I0218 15:47:14.665552 4968 scope.go:117] "RemoveContainer" containerID="a5f596668788929acd43efc0faf07605f488d17a5ceabc4b158d494ee696c99c" Feb 18 15:47:14 crc kubenswrapper[4968]: E0218 15:47:14.665804 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xnhwb_openshift-machine-config-operator(f9bae90c-908f-40fd-8373-4bf7f9aaede6)\"" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" Feb 18 15:47:27 crc kubenswrapper[4968]: I0218 15:47:27.231104 4968 scope.go:117] "RemoveContainer" containerID="a5f596668788929acd43efc0faf07605f488d17a5ceabc4b158d494ee696c99c" Feb 18 15:47:27 crc kubenswrapper[4968]: E0218 15:47:27.231910 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xnhwb_openshift-machine-config-operator(f9bae90c-908f-40fd-8373-4bf7f9aaede6)\"" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" Feb 18 15:47:40 crc kubenswrapper[4968]: I0218 15:47:40.230404 4968 scope.go:117] "RemoveContainer" containerID="a5f596668788929acd43efc0faf07605f488d17a5ceabc4b158d494ee696c99c" Feb 18 15:47:40 crc kubenswrapper[4968]: E0218 15:47:40.231114 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xnhwb_openshift-machine-config-operator(f9bae90c-908f-40fd-8373-4bf7f9aaede6)\"" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" Feb 18 15:47:51 crc kubenswrapper[4968]: I0218 15:47:51.238045 4968 scope.go:117] "RemoveContainer" containerID="a5f596668788929acd43efc0faf07605f488d17a5ceabc4b158d494ee696c99c" Feb 18 15:47:51 crc kubenswrapper[4968]: E0218 15:47:51.240380 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xnhwb_openshift-machine-config-operator(f9bae90c-908f-40fd-8373-4bf7f9aaede6)\"" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" Feb 18 15:48:03 crc kubenswrapper[4968]: I0218 15:48:03.296561 4968 scope.go:117] "RemoveContainer" containerID="a5f596668788929acd43efc0faf07605f488d17a5ceabc4b158d494ee696c99c" Feb 18 15:48:03 crc kubenswrapper[4968]: E0218 15:48:03.297978 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xnhwb_openshift-machine-config-operator(f9bae90c-908f-40fd-8373-4bf7f9aaede6)\"" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" Feb 18 15:48:14 crc kubenswrapper[4968]: I0218 15:48:14.231080 4968 scope.go:117] "RemoveContainer" containerID="a5f596668788929acd43efc0faf07605f488d17a5ceabc4b158d494ee696c99c" Feb 18 15:48:14 crc kubenswrapper[4968]: E0218 15:48:14.231896 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xnhwb_openshift-machine-config-operator(f9bae90c-908f-40fd-8373-4bf7f9aaede6)\"" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" Feb 18 15:48:25 crc kubenswrapper[4968]: I0218 15:48:25.230631 4968 scope.go:117] "RemoveContainer" containerID="a5f596668788929acd43efc0faf07605f488d17a5ceabc4b158d494ee696c99c" Feb 18 15:48:25 crc kubenswrapper[4968]: E0218 15:48:25.231857 4968 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xnhwb_openshift-machine-config-operator(f9bae90c-908f-40fd-8373-4bf7f9aaede6)\"" pod="openshift-machine-config-operator/machine-config-daemon-xnhwb" podUID="f9bae90c-908f-40fd-8373-4bf7f9aaede6" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515145357532024457 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015145357533017375 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015145354015016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015145354015015460 5ustar corecore